Unnamed: 0
int64 0
7.24k
| id
int64 1
7.28k
| raw_text
stringlengths 9
124k
| vw_text
stringlengths 12
15k
|
---|---|---|---|
1,800 | 2,635 | Heuristics for Ordering Cue Search in
Decision Making
Peter M. Todd
Anja Dieckmann
Center for Adaptive Behavior and Cognition
MPI for Human Development
Lentzeallee 94, 14195 Berlin, Germany
[email protected]
[email protected]
Abstract
Simple lexicographic decision heuristics that consider cues one at a
time in a particular order and stop searching for cues as soon as a
decision can be made have been shown to be both accurate and
frugal in their use of information. But much of the simplicity and
success of these heuristics comes from using an appropriate cue
order. For instance, the Take The Best heuristic uses validity order
for cues, which requires considerable computation, potentially
undermining the computational advantages of the simple decision
mechanism. But many cue orders can achieve good decision
performance, and studies of sequential search for data records have
proposed a number of simple ordering rules that may be of use in
constructing appropriate decision cue orders as well. Here we
consider a range of simple cue ordering mechanisms, including
tallying, swapping, and move-to-front rules, and show that they can
find cue orders that lead to reasonable accuracy and considerable
frugality when used with lexicographic decision heuristics.
1
O ne -Re ason De c i si on M aki ng and O r de r e d Se ar c h
How do we know what information to consider when making a decision? Imagine
the problem of deciding which of two objects or options is greater along some
criterion, such as which of two cities is larger. We may know various facts about
each city, such as whether they have a major sports team or a university or airport.
To decide between them, we could weight and sum all the cues we know, or we
could use a simpler lexicographic rule to look at one cue at a time in a particular
order until we find a cue that discriminates between the options and indicates a
choice [1]. Such lexicographic rules are used by people in a variety of decision
tasks [2]-[4], and have been shown to be both accurate in their inferences and frugal
in the amount of information they consider before making a decision. For instance,
Gigerenzer and colleagues [5] demonstrated the surprising performance of several
decision heuristics that stop information search as soon as one discriminating cue is
found; because only that cue is used to make the decision, and no integration of
information is involved, they called these heuristics ?one-reason? decision
mechanisms. Given some set of cues that can be looked up to make the decision,
these heuristics differ mainly in the search rule that determines the order in which
the information is searched. But then the question of what information to consider
becomes, how are these search orders determined?
Particular cue orders make a difference, as has been shown in research on the Take
The Best heuristic (TTB) [6], [7]. TTB consists of three building blocks. (1) Search
rule: Search through cues in the order of their validity, a measure of accuracy equal
to the proportion of correct decisions made by a cue out of all the times that cue
discriminates between pairs of options. (2) Stopping rule: Stop search as soon as
one cue is found that discriminates between the two options. (3) Decision rule:
Select the option to which the discriminating cue points, that is, the option that has
the cue value associated with higher criterion values.
The performance of TTB has been tested on several real-world data sets, ranging
from professors? salaries to fish fertility [8], in cross-validation comparisons with
other more complex strategies. Across 20 data sets, TTB used on average only a
third of the available cues (2.4 out of 7.7), yet still outperformed multiple linear
regression in generalization accuracy (71% vs. 68%). The even simpler Minimalist
heuristic, which searches through available cues in a random order, was more frugal
(using 2.2 cues on average), yet still achieved 65% accuracy. But the fact that the
accuracy of Minimalist lagged behind TTB by 6 percentage points indicates that
part of the secret of TTB?s success lies in its ordered search. Moreover, in
laboratory experiments [3], [4], [9], people using lexicographic decision strategies
have been shown to employ cue orders based on the cues? validities or a
combination of validity and discrimination rate (proportion of decision pairs on
which a cue discriminates between the two options).
Thus, the cue order used by a lexicographic decision mechanism can make a
considerable difference in accuracy; the same holds true for frugality, as we will
see. But constructing an exact validity order, as used by Take The Best, takes
considerable information and computation [10]. If there are N known objects to
make decisions over, and C cues known for each object, then each of the C cues
must be evaluated for whether it discriminates correctly (counting up R right
decisions), incorrectly (W wrong decisions), or does not discriminate between each
of the N?(N-1)/2 possible object pairs, yielding C?N?(N-1)/2 checks to perform to
gather the information needed to compute cue validities (v = R/(R+W)) in this
domain. But a decision maker typically does not know all of the objects to be
decided upon, nor even all the cue values for those objects, ahead of time?is there
any simpler way to find an accurate and frugal cue order?
In this paper, we address this question through simulation-based comparison of a
variety of simple cue-order-learning rules. Hope comes from two directions: first,
there are many cue orders besides the exact validity ordering that can yield good
performance; and second, research in computer science has demonstrated the
efficacy of a range of simple ordering rules for a closely related search problem.
Consequently, we find that simple mechanisms at the cue-order-learning stage can
enable simple mechanisms at the decision stage, such as lexicographic one-reason
decision heuristics, to perform well.
2
Si mpl e appr oac he s to c onstr uc ti ng c ue s e ar c h or de r s
To compare different cue ordering rules, we evaluate the performance of different cue
orders when used by a one-reason decision heuristic within a particular well-studied
sample domain: large German cities, compared on the criterion of population size using 9
cues ranging from having a university to the presence of an intercity train line [6], [7].
Examining this domain makes it clear that there are many good possible cue orders.
When used with one-reason stopping and decision building blocks, the mean accuracy of
the 362,880 (9!) cue orders is 70%, equivalent to the performance expected from
Minimalist. The accuracy of the validity order, 74.2%, falls toward the upper end of the
accuracy range (62-75.8%), but there are still 7421 cue orders that do better than the
validity order. The frugality of the search orders ranges from 2.53 cues per decision to
4.67, with a mean of 3.34 corresponding to using Minimalist; TTB has a frugality of 4.23,
implying that most orders are more frugal. Thus, there are many accurate and frugal cue
orders that could be found?a satisficing decision maker not requiring optimal
performance need only land on one.
An ordering problem of this kind has been studied in computer science for nearly four
decades, and can provide us with a set of potential heuristics to test. Consider the case of
a set of data records arranged in a list, each of which will be required during a set of
retrievals with a particular probability pi. On each retrieval, a key is given (e.g. a record?s
title) and the list is searched from the front to the end until the desired record, matching
that key, is found. The goal is to minimize the mean search time for accessing the
records in this list, for which the optimal ordering is in decreasing order of pi. But if
these retrieval probabilities are not known ahead of time, how can the list be ordered after
each successive retrieval to achieve fast access? This is the problem of self-organizing
sequential search [11], [12].
A variety of simple sequential search heuristics have been proposed for this problem,
centering on three main approaches: (1) transpose, in which a retrieved record is moved
one position closer to the front of the list (i.e., swapping with the record in front of it); (2)
move-to-front (MTF), in which a retrieved record is put at the front of the list, and all
other records remain in the same relative order; and (3) count, in which a tally is kept of
the number of times each record is retrieved, and the list is reordered in decreasing order
of this tally after each retrieval. Because count rules require storing additional
information, more attention has focused on the memory-free transposition and MTF
rules. Analytic and simulation results (reviewed in [12]) have shown that while
transposition rules can come closer to the optimal order asymptotically, in the short run
MTF rules converge more quickly (as can count rules). This may make MTF (and count)
rules more appealing as models of cue order learning by humans facing small numbers of
decision trials. Furthermore, MTF rules are more responsive to local structure in the
environment (e.g., clumped retrievals over time of a few records), and transposition can
result in very poor performance under some circumstances (e.g., when neighboring pairs
of ?popular? records get trapped at the end of the list by repeatedly swapping places).
It is important to note that there are important differences between the selforganizing sequential search problem and the cue-ordering problem we address
here. In particular, when a record is sought that matches a particular key, search
proceeds until the correct record is found. In contrast, when a decision is made
lexicographically and the list of cues is searched through, there is no one ?correct?
cue to find?each cue may or may not discriminate (allow a decision to be made).
Furthermore, once a discriminating cue is found, it may not even make the right
decision. Thus, given feedback about whether a decision was right or wrong, a
discriminating cue could potentially be moved up or down in the ordered list. This
dissociation between making a decision or not (based on the cue discrimination
rates), and making a right or wrong decision (based on the cue validities), means
that there are two ordering criteria in this problem?frugality and accuracy?as
opposed to the single order?search time?for records based on their retrieval
probability pi . Because record search time corresponds to cue frugality, the
heuristics that work well for the self-organizing sequential search task are likely to
produce orders that emphasize frugality (reflecting cue discrimination rates) over
accuracy in the cue-ordering task. Nonetheless, these heuristics offer a useful
starting point for exploring cue-ordering rules.
2.1 The cue-ordering rules
We focus on search order construction processes that are psychologically plausible by
being frugal both in terms of information storage and in terms of computation. The
decision situation we explore is different from the one assumed by Juslin and Persson
[10] who strongly differentiate learning about objects from later making decisions about
them. Instead we assume a learning-while-doing situation, consisting of tasks that have
to be done repeatedly with feedback after each trial about the adequacy of one?s decision.
For instance, we can observe on multiple occasions which of two supermarket checkout
lines, the one we have chosen or (more likely) another one, is faster, and associate this
outcome with cues including the lines? lengths and the ages of their respective cashiers.
In such situations, decision makers can learn about the differential usefulness of cues for
solving the task via the feedback received over time.
We compare several explicitly defined ordering rules that construct cue orders for
use by lexicographic decision mechanisms applied to a particular probabilistic
inference task: forced choice paired comparison, in which a decision maker has to
infer which of two objects, each described by a set of binary cues, is ?bigger? on a
criterion?just the task for which TTB was formulated. After an inference has been
made, feedback is given about whether a decision was right or wrong. Therefore,
the order-learning algorithm has information about which cues were looked up,
whether a cue discriminated, and whether a discriminating cue led to the right or
wrong decision. The rules we propose differ in which pieces of information they
use and how they use them. We classify the learning rules based on their memory
requirement?high versus low?and their computational requirements in terms of
full or partial reordering (see Table 1).
Table 1: Learning rules classified by memory and computational requirements
High memory load,
complete reordering
High memory load,
local reordering
Low memory load,
local reordering
Validity: reorders cues
based on their
current validity
Tally swap: moves
cue up (down) one
position if it has
made a correct
(incorrect) decision
if its tally of correct
minus incorrect
decisions is ? (?)
than that of next
higher (lower) cue
Simple swap: moves
cue up one position
after correct decision,
and down after an
incorrect decision
Tally: reorders cues
by number of
correct minus
incorrect decisions
made so far
Associative/delta rule:
reorders cues by
learned association
strength
Move-to-front (2 forms):
Take The Last (TTL):
moves discriminating
cue to front
TTL-correct: moves
cue to front only if it
correctly discriminates
The validity rule, a type of count rule, is the most demanding of the rules we
consider in terms of both memory requirements and computational complexity. It
keeps a count of all discriminations made by a cue so far (in all the times that the
cue was looked up) and a separate count of all the correct discriminations.
Therefore, memory load is comparatively high. The validity of each cue is
determined by dividing its current correct discrimination count by its total
discrimination count. Based on these values computed after each decision, the rule
reorders the whole set of cues from highest to lowest validity.
The tally rule only keeps one count per cue, storing the number of correct decisions
made by that cue so far minus the number of incorrect decisions. If a cue
discriminates correctly on a given trial, one point is added to its tally, if it leads to
an incorrect decision, one point is subtracted. The tally rule is less demanding in
terms of memory and computation: Only one count is kept, no division is required.
The simple swap rule uses the transposition rather than count approach. This rule
has no memory of cue performance other than an ordered list of all cues, and just
moves a cue up one position in this list whenever it leads to a correct decision, and
down if it leads to an incorrect decision. In other words, a correctly deciding cue
swaps positions with its nearest neighbor upwards in the cue order, and an
incorrectly deciding cue swaps positions with its nearest neighbor downwards.
The tally swap rule is a hybrid of the simple swap rule and the tally rule. It keeps a
tally of correct minus incorrect discriminations per cue so far (so memory load is
high) but only locally swaps cues: When a cue makes a correct decision and its tally
is greater than or equal to that of its upward neighbor, the two cues swap positions.
When a cue makes an incorrect decision and its tally is smaller than or equal to that
of its downward neighbor, the two cues also swap positions.
We also evaluate two types of move-to-front rules. First, the Take The Last (TTL)
rule moves the last discriminating cue (that is, whichever cue was found to
discriminate for the current decision) to the front of the order. This is equivalent to
the Take The Last heuristic [6], [7], which uses a memory of cues that discriminated
in the past to determine cue search order for subsequent decisions. Second, TTLcorrect moves the last discriminating cue to the front of the order only if it correctly
discriminated; otherwise, the cue order remains unchanged. This rule thus takes
accuracy as well as frugality into account.
Finally, we include an associative learning rule that uses the delta rule to update cue
weights according to whether they make correct or incorrect discriminations, and
then reorders all cues in decreasing order of this weight after each decision. This
corresponds to a simple network with nine input units encoding the difference in cue
value between the two objects (A and B) being decided on (i.e., ini = -1 if
cuei(A)<cuei(B), 1 if cuei(A)>cuei(B), and 0 if cuei(A)=cuei(B) or cuei was not
checked) and with one output unit whose target value encodes the correct decision (t
= 1 if criterion(A)>criterion(B), otherwise -1), and with the weights between inputs
and output updated according to ?wi = lr ? (t - ini?wi) ? ini with learning rate lr = 0.1.
We expect this rule to behave similarly to Oliver?s rule initially (moving a cue to
the front of the list by giving it the largest weight when weights are small) and to
swap later on (moving cues only a short distance once weights are larger).
3
Si mul ati on Study of Si mpl e O r de r i ng Rul e s
To test the performance of these order learning rules, we use the German cities data set
[6], [7], consisting of the 83 largest-population German cities (those with more than
100,000 inhabitants), described on 9 cues that give some information about population
size. Discrimination rate and validity of the cues are negatively correlated (r = -.47). We
present results averaged over 10,000 learning trials for each rule, starting from random
initial cue orders. Each trial consisted of 100 decisions between randomly selected
decision pairs. For each decision, the current cue order was used to look up cues until a
discriminating cue was found, which was used to make the decision (employing a onereason or lexicographic decision strategy). After each decision, the cue order was
updated using the particular order-learning rule. We start by considering the cumulative
accuracies (i.e., online or amortized performance?[12]) of the rules, defined as the total
percentage of correct decisions made so far at any point in the learning process. The
contrasting measure of offline accuracy?how well the current learned cue order would
do if it were applied to the entire test set?will be subsequently reported (see Figure 1).
For all but the move-to-front rules, cumulative accuracies soon rise above that of the
Minimalist heuristic (proportion correct = .70) which looks up cues in random order and
thus serves as a lower benchmark. However, at least throughout the first 100 decisions,
cumulative accuracies stay well below the (offline) accuracy that would be achieved by
using TTB for all decisions (proportion correct = .74), looking up cues in the true order of
their ecological validities. Except for the move-to-front rules, whose cumulative
accuracies are very close to Minimalist (mean proportion correct in 100 decisions: TTL:
.701; TTL-correct: .704), all learning rules perform on a surprisingly similar level, with
less than one percentage point difference in favor of the most demanding rule (i.e., delta
rule: .719) compared to the least (i.e., simple swap: .711; for comparison: tally swap:
.715; tally: .716; validity learning rule: .719). Offline accuracies are slightly higher, again
with the exception of the move to front rules (TTL: .699; TTL-correct: .702; simple
swap: .714; tally swap: .719; tally: .721; validity learning rule: .724; delta rule: .725; see
Figure 1). In longer runs (10,000 decisions) the validity learning rule is able to converge
on TTB?s accuracy, but the tally rule?s performance changes little (to .73).
Figure 1: Mean offline accuracy of
order learning rules
Figure 2: Mean offline frugality of
order learning rules
All learning rules are, however, more frugal than TTB, and even more frugal than
Minimalist, both in terms of online as well as offline frugality. Let us focus on their
offline frugality (see Figure 2): On average, the rules look up fewer cues than Minimalist
before reaching a decision. There is little difference between the associative rule, the
tallying rules and the swapping rules (mean number of cues looked up in 100 decisions:
delta rule: 3.20; validity learning rule: 3.21; tally: 3.01; tally swap: 3.04; simple swap:
3.13). Most frugal are the two move-to front rules (TTL-correct: 2.87; TTL: 2.83).
Consistent with this finding, all of the learning rules lead to cue orders that show positive
correlations with the discrimination rate cue order (reaching the following values after
100 decisions: validity learning rule: r = .18; tally: r = .29; tally swap: r = .24; simple
swap: r = .18; TTL-correct: r = .48; TTL: r = .56). This means that cues that often lead to
discriminations are more likely to end up in the first positions of the order. This is
especially true for the move-to-front rules. In contrast, the cue orders resulting from all
learning rules but the validity learning rule do not correlate or correlate negatively with
the validity cue order, and even the correlations of the cue orders resulting from the
validity learning rule after 100 decisions only reach an average r = .12.
But why would the discrimination rates of cues exert more of a pull on cue order than
validity, even when the validity learning rule is applied? As mentioned earlier, this is
what we would expect for the move-to-front rules, but it was unexpected for the other
rules. Part of the explanation comes from the fact that in the city data set we used for the
simulations, validity and discrimination rate of cues are negatively correlated. Having a
low discrimination rate means that a cue has little chance to be used and hence to
demonstrate its high validity. Whatever learning rule is used, if such a cue is displaced
downward to the lower end of the order by other cues, it may have few chances to escape
to the higher ranks where it belongs. The problem is that when a decision pair is finally
encountered for which that cue would lead to a correct decision, it is unlikely to be
checked because other, more discriminating although less valid, cues are looked up
before and already bring about a decision. Thus, because one-reason decision making is
intertwined with the learning mechanism and so influences which cues can be learned
about, what mainly makes a cue come early in the order is producing a high number of
correct decisions and not so much a high ratio of correct discriminations to total
discriminations regardless of base rates.
This argument indicates that performance may differ in environments where cue
validities and discrimination rates correlate positively. We tested the learning rules on
one such data set (r=.52) of mammal species life expectancies, predicted from 9 cues. It
also differs from the cities environment with a greater difference between TTB?s and
Minimalist?s performance (6.5 vs. 4 percentage points). In terms of offline accuracy, the
validity learning rule now indeed more closely approaches TTB?s accuracy after 100
decisions (.773 vs. .782)., The tally rule, in contrast, behaves very much as in the cities
environment, reaching an accuracy of .752, halfway between TTB and Minimalist
(accuracy =.716). Thus only some learning rules can profit from the positive correlation.
4
D i s c u s s i on
Most of the simpler cue order learning rules we have proposed do not fall far behind a
validity learning rule in accuracy, and although the move-to-front rules cannot beat the
accuracy achieved if cues were selected randomly, they compensate for this failure by
being highly frugal. Interestingly, the rules that do achieve higher accuracy than
Minimalist also beat random cue selection in terms of frugality.
On the other hand, all rules, even the delta rule and the validity learning rule, stay below
TTB?s accuracy across a relatively high number of decisions. But often it is necessary to
make good decisions without much experience. Therefore, learning rules should be
preferred that quickly lead to orders with good performance. The relatively complex rules
with relatively high memory requirement, i.e., the delta and the validity learning rule, but
also the tally learning rule, more quickly rise in accuracy compared the rules with lower
requirements. Especially the tally rule thus represents a good compromise between cost,
correctness and psychological plausibility considerations.
Remember that the rules based on tallies assume full memory of all correct minus
incorrect decisions made by a cue so far. But this does not make the rule implausible, at
least from a psychological perspective, even though computer scientists were reluctant to
adopt such counting approaches because of their extra memory requirements. There is
considerable evidence that people are actually very good at remembering the frequencies
of events. Hasher and Zacks [13] conclude from a wide range of studies that frequencies
are encoded in an automatic way, implying that people are sensitive to this information
without intention or special effort. Estes [14] pointed out the role frequencies play in
decision making as a shortcut for probabilities. Further, the tally rule and the tally swap
rule are comparatively simple, not having to keep track of base rates or perform divisions
as does the validity rule. From the other side, the simple swap and move to front rules
may not be much simpler, because storing a cue order may be about as demanding as
storing a set of tallies. We have run experiments (reported elsewhere) in which indeed the
tally swap rule best accounts for people?s actual processes of ordering cues.
Our goal in this paper was to explore how well simple cue-ordering rules could work in
conjunction with lexicographic decision strategies. This is important because it is
necessary to take into account the set-up costs of a heuristic in addition to its application
costs when considering the mechanism?s overall simplicity. As the example of the
validity search order of TTB shows, what is easy to apply may not necessarily be so easy
to set up. But simple rules can also be at work in the construction of a heuristic?s building
blocks. We have proposed such rules for the construction of one building block, the
search order. Simple learning rules inspired by research in computer science can enable a
one-reason decision heuristic to perform only slightly worse than if it had full knowledge
of cue validities from the very beginning. Giving up the assumption of full a priori
knowledge for the slight decrease in accuracy seems like a reasonable bargain: Through
the addition of learning rules, one-reason decision heuristics might lose some of their
appeal to decision theorists who were surprised by the performance of such simple
mechanisms compared to more complex algorithms, but they gain psychological
plausibility and so become more attractive as explanations for human decision behavior.
References
[1] Fishburn, P.C. (1974). Lexicographic orders, utilities and decision rules: A survey.
Management Science, 20, 1442-1471.
[2] Payne, J.W., Bettman, J.R., & Johnson, E.J. (1993). The adaptive decision maker. New York:
Cambridge University Press.
[3] Br?der, A. (2000). Assessing the empirical validity of the ?Take-The-Best? heuristic as a model
of human probabilistic inference. Journal of Experimental Psychology: Learning, Memory, and
Cognition, 26 (5), 1332-1346.
[4] Br?der, A. (2003). Decision making with the ?adaptive toolbox?: Influence of environmental
structure, intelligence, and working memory load. Journal of Experimental Psychology: Learning,
Memory, & Cognition, 29, 611-625.
[5] Gigerenzer, G., Todd, P.M., & The ABC Research Group (1999). Simple heuristics that make
us smart. New York: Oxford University Press.
[6] Gigerenzer, G., & Goldstein, D.G. (1996). Reasoning the fast and frugal way: Models of
bounded rationality. Psychological Review, 103 (4), 650-669.
[7] Gigerenzer, G., & Goldstein, D.G. (1999). Betting on one good reason: The Take The Best
Heuristic. In G. Gigerenzer, P.M. Todd & The ABC Research Group, Simple heuristics that make
us smart. New York: Oxford University Press.
[8] Czerlinski, J., Gigerenzer, G., & Goldstein, D.G. (1999). How good are simple heuristics? In G.
Gigerenzer, P.M. Todd & The ABC Research Group, Simple heuristics that make us smart. New
York: Oxford University Press.
[9] Newell, B.R., & Shanks, D.R. (2003). Take the best or look at the rest? Factors influencing
?one-reason? decision making. Journal of Experimental Psychology: Learning, Memory, and
Cognition, 29, 53-65.
[10] Juslin, P., & Persson, M. (2002). PROBabilities from EXemplars (PROBEX): a ?lazy?
algorithm for probabilistic inference from generic knowledge. Cognitive Science, 26, 563-607.
[11] Rivest, R. (1976). On self-organizing sequential search heuristics. Communications of the
ACM, 19(2), 63-67.
[12] Bentley, J.L. & McGeoch, C.C. (1985). Amortized analyses of self-organizing sequential
search heuristics. Communications of the ACM, 28(4), 404-411.
[13] Hasher, L., & Zacks, R.T. (1984). Automatic Processing of fundamental information: The case
of frequency of occurrence. American Psychologist, 39, 1372-1388.
[14] Estes, W.K. (1976). The cognitive side of probability learning. Psychological Review, 83, 3764.
| 2635 |@word trial:5 proportion:5 seems:1 simulation:3 dieckmann:2 mammal:1 profit:1 minus:5 initial:1 efficacy:1 interestingly:1 ati:1 past:1 current:5 surprising:1 si:4 yet:2 must:1 subsequent:1 analytic:1 zacks:2 update:1 v:3 discrimination:18 cue:142 implying:2 selected:2 fewer:1 intelligence:1 beginning:1 short:2 record:16 lr:2 transposition:4 successive:1 simpler:5 along:1 differential:1 become:1 surprised:1 incorrect:11 consists:1 secret:1 indeed:2 expected:1 behavior:2 mpg:2 nor:1 inspired:1 decreasing:3 little:3 actual:1 considering:2 becomes:1 moreover:1 bounded:1 rivest:1 lowest:1 what:5 ttl:11 kind:1 contrasting:1 finding:1 remember:1 ti:1 wrong:5 whatever:1 unit:2 producing:1 before:3 positive:2 influencing:1 scientist:1 local:3 todd:4 encoding:1 oxford:3 mtf:5 tallying:2 might:1 exert:1 studied:2 range:5 averaged:1 decided:2 block:4 differs:1 empirical:1 matching:1 word:1 intention:1 get:1 cannot:1 close:1 selection:1 put:1 storage:1 influence:2 equivalent:2 demonstrated:2 center:1 attention:1 starting:2 regardless:1 focused:1 survey:1 simplicity:2 rule:108 pull:1 population:3 searching:1 updated:2 imagine:1 construction:3 target:1 play:1 exact:2 rationality:1 us:4 associate:1 amortized:2 role:1 ordering:16 decrease:1 highest:1 mentioned:1 discriminates:7 accessing:1 environment:4 complexity:1 gigerenzer:7 solving:1 smart:3 compromise:1 reordered:1 upon:1 division:2 negatively:3 swap:22 various:1 train:1 forced:1 fast:2 outcome:1 whose:2 heuristic:30 larger:2 plausible:1 encoded:1 otherwise:2 favor:1 associative:3 online:2 differentiate:1 advantage:1 propose:1 neighboring:1 payne:1 organizing:4 achieve:3 moved:2 requirement:7 assessing:1 produce:1 object:9 mcgeoch:1 exemplar:1 nearest:2 received:1 dividing:1 predicted:1 come:5 differ:3 direction:1 closely:2 correct:28 subsequently:1 human:4 enable:2 require:1 generalization:1 exploring:1 hold:1 deciding:3 lentzeallee:1 cognition:4 minimalist:11 appr:1 major:1 sought:1 early:1 adopt:1 outperformed:1 lose:1 maker:5 title:1 sensitive:1 largest:2 correctness:1 city:8 hope:1 lexicographic:11 rather:1 reaching:3 ttb:16 conjunction:1 focus:2 rank:1 indicates:3 mainly:2 check:1 contrast:3 inference:5 stopping:2 typically:1 entire:1 unlikely:1 initially:1 germany:1 upward:1 overall:1 priori:1 development:1 integration:1 airport:1 uc:1 special:1 equal:3 once:2 construct:1 having:3 ng:3 represents:1 look:5 nearly:1 escape:1 employ:1 few:2 randomly:2 consisting:2 highly:1 yielding:1 swapping:4 behind:2 accurate:4 oliver:1 closer:2 partial:1 necessary:2 experience:1 respective:1 re:1 desired:1 psychological:5 instance:3 classify:1 earlier:1 ar:2 cost:3 usefulness:1 examining:1 johnson:1 front:21 reported:2 fundamental:1 discriminating:10 stay:2 probabilistic:3 quickly:3 again:1 management:1 opposed:1 fishburn:1 worse:1 cognitive:2 american:1 account:3 potential:1 de:6 explicitly:1 piece:1 later:2 doing:1 start:1 option:7 minimize:1 accuracy:31 who:2 dissociation:1 yield:1 classified:1 implausible:1 reach:1 whenever:1 checked:2 centering:1 failure:1 nonetheless:1 colleague:1 frequency:4 involved:1 associated:1 stop:3 gain:1 popular:1 reluctant:1 knowledge:3 actually:1 reflecting:1 goldstein:3 higher:5 arranged:1 evaluated:1 done:1 strongly:1 though:1 furthermore:2 just:2 stage:2 until:4 correlation:3 hand:1 working:1 bentley:1 building:4 validity:38 requiring:1 true:3 consisted:1 hence:1 laboratory:1 attractive:1 during:1 ue:1 self:4 aki:1 mpi:1 criterion:7 occasion:1 mpl:2 ini:3 complete:1 demonstrate:1 bring:1 upwards:1 reasoning:1 ranging:2 consideration:1 behaves:1 discriminated:3 association:1 he:1 slight:1 theorist:1 cambridge:1 automatic:2 similarly:1 pointed:1 clumped:1 had:1 moving:2 access:1 longer:1 base:2 retrieved:3 perspective:1 belongs:1 ecological:1 binary:1 success:2 life:1 der:2 mpib:2 greater:3 additional:1 remembering:1 converge:2 determine:1 multiple:2 full:4 infer:1 inhabitant:1 match:1 lexicographically:1 faster:1 cross:1 offer:1 retrieval:7 compensate:1 plausibility:2 bigger:1 paired:1 oac:1 regression:1 circumstance:1 checkout:1 psychologically:1 achieved:3 addition:2 extra:1 salary:1 rest:1 adequacy:1 counting:2 presence:1 easy:2 variety:3 psychology:3 juslin:2 br:2 whether:7 utility:1 effort:1 peter:1 york:4 nine:1 repeatedly:2 useful:1 se:1 clear:1 selforganizing:1 amount:1 locally:1 percentage:4 fish:1 trapped:1 delta:7 correctly:5 per:3 track:1 intertwined:1 group:3 key:3 four:1 kept:2 asymptotically:1 halfway:1 sum:1 run:3 place:1 throughout:1 reasonable:2 decide:1 decision:95 shank:1 bargain:1 encountered:1 strength:1 ahead:2 encodes:1 argument:1 relatively:3 betting:1 according:2 combination:1 poor:1 across:2 remain:1 smaller:1 slightly:2 wi:2 appealing:1 making:10 psychologist:1 remains:1 german:3 mechanism:10 count:12 needed:1 know:4 whichever:1 end:5 serf:1 available:2 apply:1 observe:1 appropriate:2 generic:1 undermining:1 occurrence:1 responsive:1 subtracted:1 include:1 estes:2 giving:2 especially:2 comparatively:2 unchanged:1 move:19 question:2 added:1 looked:5 already:1 strategy:4 distance:1 separate:1 berlin:3 reason:9 toward:1 besides:1 length:1 ratio:1 potentially:2 rise:2 lagged:1 perform:5 upper:1 displaced:1 benchmark:1 behave:1 incorrectly:2 beat:2 situation:3 looking:1 team:1 communication:2 frugal:12 pair:6 required:2 toolbox:1 learned:3 address:2 able:1 proceeds:1 below:2 including:2 memory:19 explanation:2 event:1 demanding:4 hybrid:1 ne:1 satisficing:1 fertility:1 supermarket:1 review:2 relative:1 reordering:4 expect:2 facing:1 versus:1 age:1 validation:1 gather:1 consistent:1 storing:4 pi:3 land:1 intercity:1 elsewhere:1 surprisingly:1 last:5 soon:4 transpose:1 free:1 offline:8 side:2 allow:1 fall:2 neighbor:4 wide:1 feedback:4 world:1 cumulative:4 valid:1 made:11 adaptive:3 expectancy:1 far:7 employing:1 correlate:3 emphasize:1 preferred:1 keep:4 rul:1 assumed:1 reorder:5 conclude:1 search:26 decade:1 why:1 reviewed:1 table:2 learn:1 complex:3 necessarily:1 constructing:2 domain:3 main:1 whole:1 mul:1 positively:1 downwards:1 position:9 tally:30 lie:1 third:1 down:4 load:6 list:13 appeal:1 evidence:1 sequential:7 downward:2 czerlinski:1 led:1 likely:3 explore:2 lazy:1 unexpected:1 ordered:4 sport:1 corresponds:2 newell:1 determines:1 chance:2 environmental:1 abc:3 acm:2 goal:2 formulated:1 consequently:1 professor:1 considerable:5 change:1 shortcut:1 determined:2 except:1 called:1 total:3 discriminate:3 specie:1 experimental:3 exception:1 select:1 people:5 searched:3 evaluate:2 tested:2 correlated:2 |
1,801 | 2,636 | Analysis of a greedy active learning strategy
Sanjoy Dasgupta?
University of California, San Diego
[email protected]
Abstract
We abstract out the core search problem of active learning schemes, to
better understand the extent to which adaptive labeling can improve sample complexity. We give various upper and lower bounds on the number
of labels which need to be queried, and we prove that a popular greedy
active learning rule is approximately as good as any other strategy for
minimizing this number of labels.
1
Introduction
An increasingly common phenomenon in classification tasks is that unlabeled data is abundant, whereas labels are considerably harder to come by. Genome sequencing projects, for
instance, are producing vast numbers of peptide sequences, but reliably labeling even one
of these with structural information requires time and close attention.
This distinction between labeled and unlabeled data is not captured in standard models like
the PAC framework, and has motivated the field of active learning, in which the learner
is able to ask for the labels of specific points, but is charged for each label. These query
points are typically chosen from an unlabeled data set, a practice called pool-based learning
[10]. There has also been some work on creating query points synthetically, including a
rich body of theoretical results [1, 2], but this approach suffers from two problems: first,
from a practical viewpoint, the queries thus produced can be quite unnatural and therefore
bewildering for a human to classify [3]; second, since these queries are not picked from the
underlying data distribution, they might have limited value in terms of generalization. In
this paper, we focus on pool-based learning.
We are interested in active learning with generalization guarantees. Suppose the hypothesis
class has VC dimension d and we want a classifier whose error rate on distribution P over
the joint (input, label) space, is less than > 0. The theory tells us that in a supervised
setting, we need some m = m(, d) labeled points drawn from P (for a fixed level of
confidence, which we will henceforth ignore). Can we get away with substantially fewer
than m labels if we are given unlabeled points from P and are able to adaptively choose
which points to label? How much fewer, and what querying strategies should we follow?
Here is a toy example illustrating the potential of active learning. Suppose the data lie on
the real line, and the classifiers are simple thresholding functions, H = {hw : w ? R}:
hw (x) = 1(x ? w).
VC theory tells us that if the underlying distribution P can be classified perfectly by some
hypothesis in H (called the realizable case), then it is enough to draw m = O(1/) random
labeled examples from P , and to return any classifier consistent with them. But suppose
we instead draw m unlabeled samples from P . If we lay these points down on the line,
their hidden labels are a sequence of 0?s followed by a sequence of 1?s, and the goal is to
discover the point w at which the transition occurs. This can be accomplished with a simple
binary search which asks for just log m labels. Thus active learning gives us an exponential
improvement in the number of labels needed: by adaptively querying log m labels, we can
automatically infer the rest of them.
Generalized binary search?
So far we have only looked at an extremely simple learning problem. For more complicated
hypothesis classes H, is a sort of a generalized binary search possible? What would the
search space look like? For supervised learning, in the realizable case, the usual bounds
specify a sample complexity of (very roughly) m ? d/ labeled points if the target error
rate is . So let?s pick this many unlabeled points, and then try to find a hypothesis consistent with all the hidden labels by adaptively querying just a few of them. We know via
Sauer?s lemma that H can classify these m points (considered jointly) in at most O(m d )
different ways ? in effect, the size of H is reduced to O(md ). This finite set is the effective
b (In the 1-d example, H
b has size m + 1, corresponding to the intervals
hypothesis class H.
into which the points xi split the real line.) The most we can possibly learn about the target
hypothesis, even if all labels are revealed, is to narrow it down to one of these regions. Is it
possible to pick among these O(md ) possibilities using o(m) labels? If binary search were
possible, just O(d log m) labels would be needed.
Unfortunately, we cannot hope for a generic positive result of this kind. The toy example
above is a 1-d linear separator. We show that for d ? 2, the situation is very different:
Pick any collection of m (unlabeled) points on the unit sphere in Rd ,
for d ? 2, and assume their hidden labels correspond perfectly to some
b which cannot be
linear separator. Then there are target hypotheses in H
identified without querying all the labels.
(What if the active learner is not required to identify exactly the right hypothesis, but something close to it? This and other little variations don?t help much.) Therefore, even in the
most benign situations, we cannot expect that every target hypothesis will be identifiable
using o(m) labels. To put it differently, in the worst case over target hypotheses, active
learning gives no improvement in sample complexity.
But hopefully, on average (with respect to some distribution over target hypotheses), the
number of labels needed is small. For instance, when d = 2 in the bad case above, a
b can be identified by querying just
target hypothesis chosen uniformly at random from H
O(log m) labels in expectation. This motivates the main model of this paper.
An average-case model
We will count the expected number of labels queried when the target hypothesis is chosen
b This can be interpreted as a Bayesian setting, but it is
from some distribution ? over H.
more accurate to think of ? merely as a device for averaging query counts, which has no
b
bearing on the final generalization bound. A natural choice is to make ? uniform over H.
b but H
b reflects the
Most existing active learning schemes work with H rather than H;
underlying combinatorial structure of the problem, and it can?t hurt to deal with it directly.
b for instance, if H is the set of linear
Often ? can chosen to mask the structure of H;
b is a set of convex regions of H, and ? can be made proportional to the
separators, then H
volume of each region. This makes the problem continuous rather than combinatorial.
What is the expected number of labels needed to identify a target hypothesis chosen from ??
In this average-case setting, is it always possible to get away with o(m) labels, where m is
the sample complexity of the supervised learning problem as defined above? We show that
the answer, once again, is sadly no. Thus the benefit of active learning is really a function
of the specific hypothesis class and the particular pool of unlabeled data. Depending on
these, the expected number of labels needed lies in the following range (within constants):
ideal case:
worst case:
d log m
m
perfect binary search
all labels, or randomly chosen queries
Notice the exponential gap between the top and bottom of this range. Is there some simple querying strategy which always achieves close to the minimum (expected) number of
labels, whatever this minimum number might be?
Our main result is that this property holds for a variant of a popular greedy scheme: always
ask for the label which most evenly divides the current effective version space weighted by
?. This doesn?t necessarily minimize the number of queries, just as a greedy decision tree
algorithm need not produce trees of minimum size. However:
b the expected number of labels needed by this
When ? is uniform over H,
b times that of any other strategy.
greedy strategy is at most O(ln |H|)
We also give a bound for arbitrary ?, and show corresponding lower bounds in both the
uniform and non-uniform cases.
Variants of this greedy scheme underlie many active learning heuristics, and are often described as optimal in the literature. This is the first rigorous validation of the scheme in
b = O(d log m),
a general setting. The performance guarantee is significant: recall log |H|
the minimum number of queries possible.
2
Preliminaries
Let X be the input space, Y = {0, 1} the space of labels, and P an unknown underlying
distribution over X ? Y. We want to select a hypothesis (a function X ? Y) from some
class H of VC dimension d < ?, which will accurately predict labels of points in X . We
will assume that the problem is realizable, that is, there is some hypothesis in H which
gives a correct prediction on every point. Suppose that points (x1 , y1 ) . . . , (xm , ym ) are
drawn randomly from P . Standard bounds give us a function m(, d) such that if we want
a hypothesis of error ? (on P , modulo some fixed confidence level), and if m ? m(, d),
then we need only pick a hypothesis h ? H consistent with these labeled points [9].
Now suppose just the pool of unlabeled data x1 , . . . , xm is available. The possible labelings
of these points form a subset of {0, 1}m , the effective hypothesis class
b ?
H
= {(h(x1 ), . . . , h(xm )) : h ? H}.
b = O(md ). We want to pick the unique h ? H
b which is
Sauer?s lemma [9] tells us |H|
consistent with all the hidden labels, by querying just a few of them.
Any deterministic search strategy can be represented as a binary tree whose internal nodes
b We can also
are queries (?what is the xi ?s label??), and whose leaves are elements of H.
accommodate randomization ? for instance, to allow a random choice of query point ? by
letting internal nodes of the tree be random coin flips. Our main result, Theorem 3, is
unaffected by this generalization.
xm
x1
x2
L3
x3
x4
Figure 1: To identify target hypotheses like L3 , we need to see all the labels.
3
Some bad news
Claim 1 Let H be the hypothesis class of linear separators in R2 . For any set of m distinct
data points on the perimeter of the unit circle, there are always some target hypotheses in
b which cannot be identified without querying all m labels.
H
Proof. To see this, consider the following realizable labelings (Figure 1):
? Labeling L0 : all points are negative.
? Labeling Li (1 ? i ? m): all points are negative except xi .
It is impossible to distinguish these cases without seeing all the labels.1
Remark. To rephrase this example in terms of learning a linear separator with error ? ,
suppose the input distribution P (X ) is a density over the perimeter of the unit circle. No
matter what this density is, there are always target hypotheses in H which force us to ask
for ?(1/) labels: no improvement over the sample complexity of supervised learning.
In this example, the bad target hypotheses have a large imbalance in probability mass between their positive and negative regions. By adding an extra dimension and an extra point,
exactly the same example can be modified to make the bad hypotheses balanced.
Let?s return to the original 2-d case. Some hypotheses must lie at depth m in any query tree;
but what about the rest? Well, suppose for convenience that x1 , . . . , xm are in clockwise
b = {hij : 1 ? i 6= j ? m} ? {h0 , h1 }, where hij
order around the unit circle. Then H
labels xi ? ? ? xj?1 positive (if j < i it wraps around) and the remaining points negative, and
h0 , h1 are everywhere negative/positive. It is possible to construct a query tree in which
each hij lies at depth ? 2(m/|j ? i| + log |j ? i|). Thus, if the target hypothesis is chosen
b the expected number of labels queried is at most
uniformly from H,
P
1
2m
+
2(m/|j
?
i|
+
log
|j
?
i|)
= O(log m).
i6=j
m(m?1)+2
This is why we place our hopes in an average-case analysis.
1
What if the final hypothesis ? considered as a point in {0, 1}m ? doesn?t have to be exactly right,
but within Hamming distance k of the correct one? Then a similar example forces ?(m/k) queries.
4
Main result
b we will analyze search strategies according to the number
Let ? be any distribution over H;
of labels they require, averaged over target hypotheses drawn from ?. In terms of query
trees, this is the average depth of a leaf chosen according to ?. Specifically, let T be any
tree whose leaves include the support of ?. The quality of this tree is
P
P
Q(T, ?) =
b ?(h) ? leaf-depth(h).
b ?(h) ? (# labels needed for h) =
h?H
h?H
Is there always a tree of average depth o(m)? The answer, once again, is sadly no.
Claim 2 Pick any d ? 2 and any m ? 2d. There is an input space X of size m and a
hypothesis class H of VC dimension d, defined on domain X , with the following property:
b then any query tree T has Q(T, ?) ? m/8.
if ? is chosen to be uniform over H = H,
Proof. Let X consist of any m points x1 , . . . , xm , and let H consist of all hypotheses
h : X ? {0, 1} which are positive on exactly d inputs. In order to identify a particular
element h ? H, any querying method must discover exactly the d points xi on which h is
nonzero. By construction, the order in which queries are asked is irrelevant ? it might as
well be x1 , x2 , . . .. The rest is a simple probability calculation.
In our average-case model, we have seen one example in which intelligent querying results
in an exponential improvement in the number of labels required, and one in which it is
no help at all. Is there some generic scheme which always comes close to minimizing the
number of queries, whatever the minimum number might be? Here?s a natural candidate:
b be the current version space. For each
Greedy strategy. Let S ? H
unlabeled xi , let Si+ be the hypotheses which label xi positive and Si?
the ones which label it negative. Pick the xi for which these sets are most
nearly equal in ?-mass, that is, for which min{?(Si+ ), ?(Si? )} is largest.
We show this is almost as good at minimizing queries as any other strategy.
b Suppose that the optimal query tree requires
Theorem 3 Let ? be any distribution over H.
?
Q labels in expectation, for target hypotheses chosen according to ?. Then the expected
number of labels needed by the greedy strategy is at most 4Q? ln 1/(minh ?(h)).
b We also show
For the case of uniform ?, the approximation ratio is thus at most 4 ln |H|.
almost-matching lower bounds in both the uniform and non-uniform cases.
5
5.1
Analysis of the greedy active learner
Lower bounds on the greedy scheme
The greedy approach is not optimal because it doesn?t take into account the way in which a
query reshapes the search space ? specifically, the effect of a query on the quality of other
b might consist of several dense clusters, each of which permits
queries. For instance, H
rapid binary search. However, the version space must first be whittled down to one of
these subregions, and this process, though ultimately optimal, might initially be slower at
shrinking the hypothesis space than more shortsighted alternatives. A concrete example of
this type gives rise to the following lower bound.
b n of size n
Claim 4 For any n ? 16 which is a power of two, there is a concept class H
such that: under uniform ?, the optimal tree has average height at most qn = ?(log n),
but the greedy active learning strategy produces a tree of average height ?(q n ? logloglogn n ).
For non-uniform ?, the greedy scheme can deviate more substantially from optimality.
b with 2n + 1 elements and a distriClaim 5 For any n ? 2, there is a hypothesis class H
b
bution ? over H, such that: (a) ? ranges in value from 1/2 to 1/2n+1 ; (b) the optimal tree
has average depth less than 3; (c) the greedy tree has average depth at least n/2.
Proofs of these lower bounds appear in the full paper, available at the author?s website.
5.2
Upper bound
Overview. The lower bounds on the quality of a greedy learner are sobering, but things
cannot get too much worse than this. Here?s the basic argument for uniform ?: we show
that if the optimal tree T ? requires Q? queries in expectation, then some query must (again
b of ?-mass ?(1/Q? ). Therefore, the root query of
in expectation) ?cut off? a chunk of H
the greedy tree TG is at least this good (cf. Johnson?s set cover analysis [8]). Things get
trickier when we try to show that the rest of TG is also good, because although T ? uses just
Q? queries on average, it may need many more queries for certain hypotheses. Subtrees
of TG could correspond to version spaces for which more than Q? queries are needed, and
the roots of these subtrees might not cut down the version space much...
For a worst-case model, a proof of approximate optimality is known in a related context
[6]; as we saw in Claim 1, that model is trivial in our situation. The average-case model,
and especially the use of arbitrary weights ?, require more care.
Details. For want of space, we only discuss some issues that arise in proving the main
theorem, and leave the actual proof to the full paper.
The key concept we have to define is the quality of a query, and it turns out that we need this
to be monotonically decreasing, that is, it should only go down as active learning proceeds
and the version space shrinks. This rules out some natural entropy-based notions.
b and a possible next query is xj . If
Suppose we are down to some version space S ? H,
S + is the subset of S which labels xj positive, and S ? are the ones that label it negative,
then on average the probability mass (measured by ?) eliminated by xj is
?(S + )
?
?(S) ?(S )
+
?(S ? )
+
?(S) ?(S )
=
2?(S + )?(S ? )
.
?(S)
We say xj shrinks (S, ?) by this much, with the understanding that this is in expectation.
Shrinkage is easily seen to have the monotonicity property we need.
b ?) by ?, then it shrinks (S, ?) by at most ? for any S ? H.
b
Lemma 6 If xj shrinks (H,
We would expect that if the optimal tree is short, there must be at least one query which
b ?) considerably. More concretely, the definition of shrinkage seems to suggest
shrinks (H,
that if all queries provide shrinkage at most ?, and the current version space has mass ?(S),
then at least about ?(S)/? more queries are needed. This isn?t entirely true, because of a
second effect: if |S| = 2, then we need just one query, regardless of ?(S).
Roughly speaking, when there are lots of hypotheses with significant mass left in S, the first
effect dominates; thereafter the second takes over. To smoothly incorporate both effects,
we use the
Pnotion of collision probability. For a distribution ? over support Z, this is
CP(?) = z?Z ?(z)2 , the chance that two random draws from ? are identical.
b ?) by at most ? > 0. Pick any S ? H,
b
Lemma 7 Suppose every query shrinks (H,
and any query tree T whose leaves include S. If ?S is the restriction of ? to S (that is,
?S (h) = ?(h)/?(S) for h ? S), then Q(T, ?S ) ? (1 ? CP(?S )) ? ?(S)/?.
b and any tree T whose leaves include all of S. Then there
Corollary 8 Pick any S ? H
must exist a query which shrinks (S, ?S ) by at least (1 ? CP(?S ))/Q(T, ?S ).
b is such that ?S has small collision probability, some
So if the current version space S ? H
query must split off a sizeable chunk of S. This can form the basis of a proof by induction.
But what if CP(?S ) is large, say greater than 1/2? In this case, the mass of some particular
hypothesis h0 ? S exceeds that of all the others combined, and S could shrink by just
an insignificant amount during the subsequent greedy query, or even during the next few
iterations of greedy queries. It turns out, however, that within roughly the number of iterations that the optimal tree needs for target h0 , the greedy procedure will either reject h0 or
identify it as the target. If it is rejected, then by that time S will have shrunk considerably.
By combining the two cases for CP(?S ), we get the following lemma, which is proved in
the full paper and yields our main theorem as an immediate consequence.
Lemma 9 Let T ? denote any particular query tree for ?, and let T be the greedilyb which corresponds to a subtree TS of T ,
constructed query tree. For any S ? H
?(S)
Q(TS , ?S ) ? 4Q(T ? , ?S ) ln
.
minh?S ?(h)
6
Related work and promising directions
Rather than attempting to summarize the wide range of proposed active learning methods,
for instance [5, 7, 10, 13, 14], we will discuss three basic techniques upon which they rely.
Greedy search. This is the technique we have abstracted and rigorously validated in this
paper. It is the foundation of most of the schemes cited above. Algorithmically, the main
problem is that the query selection rule is not immediately tractable, so approximations are
b consists of convex sets, and if ? is chosen to be pronecessary. For linear separators, H
portional to volume, query selection involves estimating volumes of convex regions, which
is tractable but (using present techniques) inconvenient. Tong and Koller [13] investigate
margin-based approximations which are efficiently computable using SVM technology.
Opportunistic priors. This is a trick in which the learner takes a look at the unlabeled
b leads to standard
data and then places bets on hypotheses. A uniform bet over all of H
generalization bounds. But if the algorithm places more weight on certain hypotheses (for
instance, those with large margin), then its final error bound is excellent if it guessed right,
and worse-than-usual if it guessed wrong. This technique is not specific to active learning,
and has been analyzed elsewhere (eg. [12]). One interesting line of work investigates a
flexible family of priors specified by pairwise similarities between data points, eg. [14].
Bayesian assumptions. In our analysis, although ? can be seen as some sort of prior belief,
there is no assumption that nature shares this belief; in particular, the generalization bound
does not depend on it. A Bayesian assumption has an immediate benefit for active learning:
if at any stage the remaining version space (weighted by prior ?) is largely in agreement
on the unlabeled data, it is legitimate to stop and output one of these remaining hypotheses
[7]. In a non-Bayesian setting this is not legitimate.
When the hypothesis class consists of probabilistic classifiers, the Bayesian assumption
has also been used in another way: to approximate the greedy selection rule using the
MAP estimate instead of an expensive summation over the posterior (eg. [11]).
In terms of theoretical results, another work which considers the tradeoff between labels
and generalization error is [7], in which a greedy scheme, realized using sampling, is analyzed in a Bayesian setting. The authors show that it is possible to achieve an exponential
improvement in the number of labels needed to learn linear separators, when both data and
target hypothesis are chosen uniformly from the unit sphere. It is an intriguing question
whether this holds for more general data distributions.
Other directions. We have looked at the case where the acceptable error rate is fixed and
the goal is to minimize the number of queries. What about fixing the number of queries and
asking for the best (average) error rate possible? In other words, the query tree has a fixed
b Treating each
depth, and each leaf is annotated with its remaining version space S ? H.
element of S as a point in {0, 1}m (its predictions on the pool of data), the error at this leaf
depends on the Hamming diameter of S. What is a good querying strategy for producing
low-diameter leaves?
The most widely-used classifiers are perhaps linear separators. Existing active learning
b an arrangement of hyperplanes [4].
schemes ignore the rich algebraic structure of H,
Acknowledgements. I am very grateful to the anonymous NIPS reviewers for their careful
and detailed feedback.
References
[1] D. Angluin. Queries and concept learning. Machine Learning, 2:319?342, 1988.
[2] D. Angluin. Queries revisited. Proceedings of the Twelfth International Conference on Algorithmic Learning Theory, pages 12?31, 2001.
[3] E.B. Baum and K. Lang. Query learning can work poorly when a human oracle is used. International Joint Conference on Neural Networks, 1992.
[4] A. Bjorner, M. Las Vergnas, B. Sturmfels, N. White, and G. Ziegler. Oriented matroids. Cambridge University Press, 1999.
[5] D. Cohn, Z. Ghahramani, and M. Jordan. Active learning with statistical models. Journal of
Artificial Intelligence Research, 4:129?145, 1996.
[6] S. Dasgupta, P.M. Long, and W.S. Lee. A theoretical analysis of query selection for collaborative filtering. Machine Learning, 51:283?298, 2003.
[7] Y. Freund, S. Seung, E. Shamir, and N. Tishby. Selective sampling using the query by committee
algorithm. Machine Learning, 28:133?168, 1997.
[8] D.S. Johnson. Approximation algorithms for combinatorial problems. Journal of Computer
and System Sciences, 9:256?278, 1974.
[9] M.J. Kearns and U.V. Vazirani. An introduction to computational learning theory. MIT Press,
1993.
[10] A. McCallum and K. Nigam. Employing em and pool-based active learning for text classification. Fifteenth International Conference on Machine Learning, 1998.
[11] N. Roy and A. McCallum. Toward optimal active learning through sampling of error reduction.
Twentieth International Conference on Machine Learning, 2003.
[12] J. Shawe-Taylor, P. Bartlett, R. Williamson, and M. Anthony. Structural risk minimization over
data-dependent hierarchies. IEEE Transactions on Information Theory, 1998.
[13] S. Tong and D. Koller. Support vector machine active learning with applications to text classification. Journal of Machine Learning Research, 2001.
[14] X. Zhu, J. Lafferty, and Z. Ghahramani. Combining active learning and semi-supervised learning using gaussian fields and harmonic functions. ICML workshop, 2003.
| 2636 |@word illustrating:1 version:11 seems:1 twelfth:1 pick:9 asks:1 accommodate:1 harder:1 whittled:1 reduction:1 existing:2 current:4 si:4 lang:1 intriguing:1 must:7 subsequent:1 benign:1 treating:1 greedy:22 fewer:2 device:1 leaf:9 website:1 intelligence:1 mccallum:2 core:1 short:1 node:2 revisited:1 hyperplanes:1 height:2 constructed:1 prove:1 consists:2 pairwise:1 mask:1 expected:7 rapid:1 roughly:3 decreasing:1 automatically:1 little:1 actual:1 project:1 discover:2 underlying:4 estimating:1 mass:7 what:11 kind:1 interpreted:1 substantially:2 guarantee:2 every:3 exactly:5 classifier:5 wrong:1 whatever:2 unit:5 underlie:1 appear:1 producing:2 positive:7 consequence:1 approximately:1 might:7 limited:1 range:4 averaged:1 practical:1 unique:1 practice:1 x3:1 opportunistic:1 procedure:1 reject:1 matching:1 confidence:2 word:1 seeing:1 suggest:1 get:5 cannot:5 unlabeled:12 close:4 convenience:1 selection:4 put:1 context:1 impossible:1 risk:1 restriction:1 deterministic:1 charged:1 map:1 reviewer:1 baum:1 go:1 attention:1 regardless:1 convex:3 immediately:1 legitimate:2 rule:4 proving:1 notion:1 variation:1 hurt:1 diego:1 suppose:10 target:19 modulo:1 construction:1 shamir:1 hierarchy:1 us:1 hypothesis:44 agreement:1 trick:1 element:4 roy:1 expensive:1 lay:1 cut:2 labeled:5 bottom:1 worst:3 sadly:2 region:5 news:1 balanced:1 complexity:5 asked:1 seung:1 rigorously:1 ultimately:1 depend:1 grateful:1 upon:1 learner:5 basis:1 easily:1 joint:2 differently:1 various:1 represented:1 distinct:1 effective:3 query:52 artificial:1 labeling:4 tell:3 h0:5 quite:1 whose:6 heuristic:1 widely:1 say:2 think:1 jointly:1 final:3 sequence:3 combining:2 poorly:1 achieve:1 cluster:1 produce:2 perfect:1 leave:1 help:2 depending:1 fixing:1 measured:1 c:1 involves:1 come:2 direction:2 correct:2 annotated:1 shrunk:1 vc:4 human:2 require:2 generalization:7 really:1 preliminary:1 randomization:1 anonymous:1 summation:1 hold:2 around:2 considered:2 algorithmic:1 predict:1 claim:4 achieves:1 label:50 combinatorial:3 ziegler:1 peptide:1 saw:1 largest:1 reflects:1 weighted:2 hope:2 minimization:1 mit:1 always:7 gaussian:1 modified:1 rather:3 shrinkage:3 bet:2 corollary:1 l0:1 focus:1 validated:1 improvement:5 sequencing:1 rigorous:1 realizable:4 am:1 dependent:1 typically:1 initially:1 hidden:4 koller:2 selective:1 labelings:2 interested:1 issue:1 classification:3 among:1 flexible:1 field:2 once:2 construct:1 equal:1 eliminated:1 sampling:3 x4:1 identical:1 look:2 icml:1 nearly:1 others:1 intelligent:1 few:3 randomly:2 oriented:1 possibility:1 investigate:1 analyzed:2 perimeter:2 subtrees:2 accurate:1 sauer:2 tree:25 divide:1 taylor:1 abundant:1 circle:3 inconvenient:1 theoretical:3 instance:7 classify:2 asking:1 cover:1 trickier:1 tg:3 subset:2 uniform:12 johnson:2 too:1 tishby:1 answer:2 considerably:3 combined:1 adaptively:3 chunk:2 density:2 cited:1 international:4 probabilistic:1 off:2 lee:1 pool:6 ym:1 concrete:1 again:3 choose:1 possibly:1 henceforth:1 worse:2 creating:1 return:2 toy:2 li:1 account:1 potential:1 sizeable:1 matter:1 depends:1 try:2 picked:1 h1:2 root:2 analyze:1 lot:1 bution:1 sort:2 complicated:1 collaborative:1 minimize:2 largely:1 efficiently:1 correspond:2 identify:5 yield:1 guessed:2 bayesian:6 accurately:1 produced:1 unaffected:1 classified:1 suffers:1 definition:1 bewildering:1 proof:6 hamming:2 stop:1 proved:1 popular:2 ask:3 recall:1 supervised:5 follow:1 specify:1 though:1 shrink:8 just:10 rejected:1 stage:1 cohn:1 hopefully:1 quality:4 perhaps:1 effect:5 concept:3 true:1 nonzero:1 deal:1 eg:3 white:1 during:2 generalized:2 cp:5 harmonic:1 common:1 overview:1 volume:3 significant:2 cambridge:1 queried:3 rd:1 i6:1 shawe:1 l3:2 similarity:1 something:1 posterior:1 irrelevant:1 certain:2 binary:7 accomplished:1 captured:1 minimum:5 seen:3 care:1 greater:1 monotonically:1 clockwise:1 semi:1 full:3 infer:1 exceeds:1 calculation:1 sphere:2 long:1 prediction:2 variant:2 basic:2 expectation:5 fifteenth:1 iteration:2 whereas:1 want:5 interval:1 extra:2 rest:4 thing:2 lafferty:1 jordan:1 structural:2 synthetically:1 revealed:1 split:2 enough:1 ideal:1 xj:6 perfectly:2 identified:3 computable:1 tradeoff:1 whether:1 motivated:1 bartlett:1 unnatural:1 algebraic:1 speaking:1 remark:1 collision:2 detailed:1 amount:1 sturmfels:1 subregions:1 diameter:2 reduced:1 angluin:2 exist:1 notice:1 algorithmically:1 dasgupta:3 key:1 thereafter:1 drawn:3 vast:1 merely:1 everywhere:1 place:3 almost:2 family:1 draw:3 decision:1 acceptable:1 investigates:1 entirely:1 bound:15 followed:1 distinguish:1 identifiable:1 oracle:1 x2:2 argument:1 extremely:1 min:1 optimality:2 attempting:1 according:3 increasingly:1 em:1 ln:4 discus:2 count:2 turn:2 committee:1 needed:11 know:1 letting:1 flip:1 tractable:2 available:2 permit:1 shortsighted:1 away:2 generic:2 alternative:1 coin:1 slower:1 original:1 top:1 remaining:4 include:3 cf:1 ghahramani:2 especially:1 question:1 realized:1 occurs:1 looked:2 strategy:13 arrangement:1 usual:2 md:3 wrap:1 distance:1 evenly:1 extent:1 considers:1 trivial:1 toward:1 induction:1 ratio:1 minimizing:3 unfortunately:1 hij:3 negative:7 rise:1 reliably:1 motivates:1 unknown:1 upper:2 imbalance:1 finite:1 minh:2 t:2 immediate:2 situation:3 y1:1 ucsd:1 arbitrary:2 required:2 specified:1 rephrase:1 california:1 distinction:1 narrow:1 nip:1 able:2 proceeds:1 xm:6 summarize:1 including:1 belief:2 power:1 natural:3 force:2 rely:1 zhu:1 scheme:11 improve:1 technology:1 isn:1 deviate:1 prior:4 literature:1 understanding:1 acknowledgement:1 text:2 freund:1 expect:2 interesting:1 proportional:1 filtering:1 querying:11 validation:1 foundation:1 consistent:4 thresholding:1 viewpoint:1 share:1 elsewhere:1 allow:1 understand:1 wide:1 matroids:1 benefit:2 feedback:1 dimension:4 depth:8 transition:1 genome:1 rich:2 doesn:3 qn:1 collection:1 adaptive:1 san:1 made:1 author:2 concretely:1 far:1 employing:1 transaction:1 vazirani:1 approximate:2 ignore:2 monotonicity:1 abstracted:1 active:24 xi:8 don:1 search:12 continuous:1 why:1 promising:1 learn:2 nature:1 nigam:1 bearing:1 excellent:1 necessarily:1 separator:8 williamson:1 domain:1 anthony:1 main:7 dense:1 arise:1 body:1 x1:7 tong:2 shrinking:1 exponential:4 lie:4 candidate:1 hw:2 down:6 theorem:4 bad:4 specific:3 pac:1 r2:1 insignificant:1 svm:1 dominates:1 reshapes:1 consist:3 workshop:1 adding:1 subtree:1 margin:2 gap:1 entropy:1 smoothly:1 twentieth:1 corresponds:1 portional:1 chance:1 goal:2 careful:1 specifically:2 except:1 uniformly:3 averaging:1 lemma:6 kearns:1 called:2 sanjoy:1 la:1 select:1 internal:2 support:3 incorporate:1 phenomenon:1 |
1,802 | 2,637 | Online Bounds for Bayesian Algorithms
Sham M. Kakade
Computer and Information Science Department
University of Pennsylvania
Andrew Y. Ng
Computer Science Department
Stanford University
Abstract
We present a competitive analysis of Bayesian learning algorithms in the
online learning setting and show that many simple Bayesian algorithms
(such as Gaussian linear regression and Bayesian logistic regression) perform favorably when compared, in retrospect, to the single best model in
the model class. The analysis does not assume that the Bayesian algorithms? modeling assumptions are ?correct,? and our bounds hold even
if the data is adversarially chosen. For Gaussian linear regression (using logloss), our error bounds are comparable to the best bounds in the
online learning literature, and we also provide a lower bound showing
that Gaussian linear regression is optimal in a certain worst case sense.
We also give bounds for some widely used maximum a posteriori (MAP)
estimation algorithms, including regularized logistic regression.
1
Introduction
The last decade has seen significant progress in online learning algorithms that perform well
even in adversarial settings (e.g. the ?expert? algorithms of Cesa-Bianchi et al. (1997)). In
the online learning framework, one makes minimal assumptions on the data presented to
the learner, and the goal is to obtain good (relative) performance on arbitrary sequences. In
statistics, this philosophy has been espoused by Dawid (1984) in the prequential approach.
We study the performance of Bayesian algorithms in this adversarial setting, in which the
process generating the data is not restricted to come from the prior?data sequences may
be arbitrary. Our motivation is similar to that given in the online learning literature and the
MDL literature (see Grunwald, 2005) ?namely, that models are often chosen to balance
realism with computational tractability, and often assumptions made by the Bayesian are
not truly believed to hold (e.g. i.i.d. assumptions). Our goal is to study the performance of
Bayesian algorithms in the worst-case, where all modeling assumptions may be violated.
We consider the widely used class of generalized linear models?focusing on Gaussian
linear regression and logistic regression?and provide relative performance bounds (comparing to the best model in our model class) when the cost function is the logloss. Though
the regression problem has been studied in a competitive framework and, indeed, many
ingenious algorithms have been devised for it (e.g., Foster, 1991; Vovk, 2001; Azoury and
Warmuth, 2001) , our goal here is to study how the more widely used, and often simpler,
Bayesian algorithms fare. Our bounds for linear regression are comparable to the best
bounds in the literature (though we use the logloss as opposed to the square loss).
The competitive approach to regression started with Foster (1991), who provided competitive bounds for a variant of the ridge regression algorithm (under the square loss).
Vovk (2001) presents many competitive algorithms and provides bounds for linear regression (under the square loss) with an algorithm that differs slightly from the Bayesian one.
Azoury and Warmuth (2001) rederive Vovk?s bound with a different analysis based on
Bregman distances. Our work differs from these in that we consider Bayesian Gaussian
linear regression, while previous work typically used more complex, cleverly devised algorithms which are either variants of a MAP procedure (as in Vovk, 2001) , or that involve
other steps such as ?clipping? predictions (as in Azoury and Warmuth, 2001) . These distinctions are discussed in more detail in Section 3.1.
We should also note that when the loss function is the logloss, multiplicative weights algorithms are sometimes identical to Bayes rule with particular choices of the parameters (see
Freund and Schapire, 1999) . Furthermore, Bayesian algorithms have been used in some
online learning settings, such as the sleeping experts setting of Freund et al. (1997) and the
online boolean prediction setting of Cesa-Bianchi et al. (1998). Ng and Jordan (2001) also
analyzed an online Bayesian algorithm but assumed that the data generation process was
not too different from the model prior. To our knowledge, there have been no studies of
Bayesian generalized linear models in an adversarial online learning setting (though many
variants have been considered as discussed above).
We also examine maximum a posteriori (MAP) algorithms for both Gaussian linear regression (i.e., ridge regression) and for (regularized) logistic regression. These algorithms are
often used in practice, particularly in logistic regression where Bayesian model averaging is computationally expensive, but the MAP algorithm requires only solving a convex
problem. As expected, MAP algorithms are somewhat less competitive than full Bayesian
model averaging, though not unreasonably so.
2
Bayesian Model Averaging
We now consider the Bayesian model averaging (BMA) algorithm and give a bound on its
worst-case online loss. We start with some preliminaries. Let x ? Rn denote the inputs of
a learning problem and y ? R the outputs. Consider a model from the generalized linear
model family (see McCullagh and Nelder, 1989) , that can be written p(y|x, ?) = p(y|?T x),
where ? ? Rn are the parameters of our model (?T denotes the transpose of ?). Note that
the predicted distribution of y depends only on ?T x, which is linear in ?. For example, in
the case of Gaussian linear regression, we have
?(?T x ? y)2
1
exp
p(y|x, ?) = ?
,
(1)
2? 2
2?? 2
where ? 2 is a fixed, known constant that is not a parameter of our model. In logistic
regression, we would have
1
1
log p(y|x, ?) = y log
+
(1
?
y)
log
1
?
,
(2)
1 + exp(??T x)
1 + exp(??T x)
where we assume y ? {0, 1}.
Let S = {(x(1) , y (1) ), (x(2) , y (2) ), . . . , (x(T ) , y (T ) )} be an arbitrary sequence of examples,
possibly chosen by an adversary. We also use St to denote the subsequence consisting of
only the first t examples. We assume throughout this paper that ||x(t) || ? 1 (where || ? ||
denotes the L2 norm).
Assume that we are going to use a Bayesian algorithm to make our online predictions.
Specifically, assume that we have a Gaussian prior on the parameters:
p(?) = N (?; ~0, ? 2 In ),
where In is the n-by-n identity matrix, N (?; ?, ?) is the Gaussian density with mean ? and
covariance ?, and ? 2 > 0 is some fixed constant governing the prior variance. Also, let
Q
t
(i) (i)
i=1 p(y |x , ?) p(?)
pt (?) = p(?|St ) = R Q
t
(i) |x(i) , ?) p(?)d?
p(y
i=1
?
be the posterior distribution over ? given the first t training examples. We have that p0 (?) =
p(?) is just the prior distribution.
On iteration t, we are given the input x(t) , and our algorithm makes a prediction using the
posterior distribution over the outputs:
Z
p(y|x(t) , St?1 ) = p(y|x(t) , ?)p(?|St?1 )d?.
?
We are then given the true label y (t) , and we suffer logloss ? log p(y (t) |x(t) , St?1 ). We
define the cumulative loss of the BMA algorithm after T rounds to be
T
X
LBMA (S) =
? log p(y (t) |x(t) , St?1 ).
t=1
Importantly, note that even though the algorithm we consider is a Bayesian one, our theoretical results do not assume that the data comes from any particular probabilistic model.
In particular, the data may be chosen by an adversary.
We are interested in comparing against the loss of any ?expert? that uses some fixed parameters ? ? Rn . Define `? (t) = ? log p(y (t) |x(t) , ?), and let
T
T
X
X
L? (S) =
`? (t) =
? log p(y (t) |x(t) , ?).
t=1
t=1
Sometimes, we also wish to
R compare against distributions over experts. Given a distribution
Q over ?, define `Q (t) = ? ?Q(?) log p(y (t) |x(t) , ?)d?, and
Z
T
X
LQ (S) =
`Q (t) = Q(?)L? (S)d?.
t=1
?
This is the expected logloss incurred by a procedure that first samples some ? ? Q and
then uses this ? for all its predictions. Here, the expectation is over the random ?, not over
the sequence of examples. Note that the expectation is of the logloss, which is a different
type of averaging than in BMA, which had the expectation and the log in the reverse order.
2.1 A Useful Variational Bound
The following lemma provides a worst case bound of the loss incurred by Bayesian algorithms and will be useful for deriving our main result in the next section. A result very
similar to this (for finite model classes) is given by Freund et al. (1997). For completeness,
we prove the result here in its full generality, though our proof is similar to theirs. As usual,
R
q(?)
.
define KL(q||p) = ? q(?) log p(?)
Lemma 2.1: Let Q be any distribution over ?. Then for all sequences S
LBMA (S) ? LQ (S) + KL(Q||p0 ).
Proof: Let Y = {y (1) , . . . , y (T ) } and X = {x(1) , . . . , x(T ) }. The chain rule of conditional
probabilities implies that LBMA (S) = ? log p(Y |X) and L? (S) = ? log p(Y |X, ?). So
Z
LBMA (S) ? LQ (S) = ? log p(Y |X) + Q(?) log p(Y |X, ?)d?
?
Z
p(Y |X, ?)
=
Q(?) log
d?
p(Y |X)
?
|X,?)p0 (?)
By Bayes rule, we have that pT (?) = p(Yp(Y
. Continuing,
|X)
Z
pT (?)
=
Q(?) log
d?
p0 (?)
Z?
Z
Q(?)
Q(?)
=
Q(?) log
d? ? Q(?) log
d?
p
(?)
p
0
T (?)
?
?
= KL(Q||p0 ) ? KL(Q||pT ).
Together with the fact that KL(Q||pT ) ? 0, this proves the lemma.
2.2 An Upper Bound for Generalized Linear Models
For the theorem that we shortly present, we need one new definition. Let fy (z) =
? log p(y|?T x = z). Thus, fy(t) (?T x(t) ) = `? (t). Note that for linear regression (as
defined in Equation 1), we have that for all y
1
(3)
|fy00 (z)| = 2
?
and for logistic regression (as defined in Equation 2), we have that for y ? {0, 1}
|fy00 (z)| ? 1 .
Theorem 2.2: Suppose fy (z) is continuously differentiable. Let S be a sequence such that
||x(t) || ? 1 and such that for some constant c, |fy00(t) (z)| ? c (for all z). Then for all ?? ,
1
n
T c? 2
? 2
LBMA (S) ? L?? (S) + 2 ||? || + log 1 +
(4)
2?
2
n
The ||?? ||2 /2? 2 term can be interpreted as a penalty term from our prior. The log term is
how fast our loss could grow in comparison to the best ?? . Importantly, this extra loss is
only logarithmic in T in this adversarial setting.
This bound almost identical to those provided by Vovk (2001); Azoury and Warmuth (2001)
and Foster (1991) for the linear regression case (under the square loss); the only difference
is that in their bounds, the last term is multiplied by an upper bound on y (t) . In contrast,
we require no bound on y (t) in the Gaussian linear regression case due to the fact that we
deal with the logloss (also recall |fy00 (z)| = ?12 for all y).
Proof: We use Lemma 2.1 with Q(?) = N (?; ?? , 2 In ) being a normal distribution with
mean ?? and covariance 2 In . Here, 2 is a variational parameter that we later tune to get
the tightest possible bound. Letting H(Q) = n2 log 2?e2 be the entropy of Q, we have
?1
Z
1
1 T
KL(Q||p0 ) =
Q(?) log
exp
?
?
?
d? ? H(Q)
2? 2
(2?)n/2 |? 2 In |1/2
?
Z
1
n
= n log ? + 2
Q(?)?T ?d? ? ? n log
2? ?
2
1
n
= n log ? + 2 ||?? ||2 + n2 ? ? n log .
(5)
2?
2
To prove the result, we also need to relate the error of LQ to that of L?? . By taking a Taylor
expansion of fy (assume y ? S), we have that
(z ? z ? )2
fy (z) = fy (z ? ) + fy0 (z ? )(z ? z ? ) + fy00 (?(z))
,
2
for some appropriate function ?. Thus, if z is a random variable with mean z ? , we have
(z ? z ? )2
Ez [fy (z)] = fy (z ? ) + fy0 (z ? ) ? 0 + Ez fy00 (?(z))
2
? 2
(z ? z )
? fy (z ? ) + cEz
2
c
= fy (z ? ) + Var(z)
2
Consider a single example (x, y). We can apply the argument above with z ? = ?? T x, and
z = ?T x, where ? ? Q. Note that E[z] = z ? since Q has mean ?? . Also, Var(?T x) =
xT (2 In )x = ||x||2 2 ? 2 (because we previously assumed that ||x|| ? 1). Thus, we have
c2
E??Q [fy (?T x)] ? fy (?? T x) +
2
Since `Q (t) = E??Q [fy(t) (?T x(t) )] and `?? (t) = fy(t) (?? T x(t) ), we can sum both sides
from t = 1 to T to obtain
Tc 2
LQ (S) ? L?? (S) +
2
Putting this together with Lemma 2.1 and Equation 5, we find that
n
Tc 2
1
LBMA (S) ? L?? (S) +
+ n log ? + 2 ||?? ||2 + n2 ? ? n log .
2
2?
2
n? 2
and
simplifying,
Theorem
2.2
follows.
Finally, by choosing 2 = n+T
c? 2
2.3 A Lower Bound for Gaussian Linear Regression
The following lower bound shows that, for linear regression, no other prediction scheme is
better than Bayes in the worst case (when our penalty term is ||?? ||2 ). Here, we compare to
an arbitrary predictive distribution q(y|x(t) , St?1 ) for prediction at time t, which suffers an
instant loss `q (t) = ? log q(y (t) |x(t) , St?1 ). In the theorem, b?c denotes the floor function.
Theorem 2.3: Let L?? (S) be the loss under the Gaussian linear regression model using the
parameter ?? , and let ? 2 = ? 2 = 1. For any set of predictive distributions q(y|x(t) , St?1 ),
there exists an S with ||x(t) || ? 1 such that
T
X
1 ? 2
n
T
? (S) +
`q (t) ? inf
(L
||?
||
)
+
log
1
+
?
?
?
2
2
n
t=1
Proof: (sketch) If n = 1 and if S is such that x(t) = 1, one can show the equality:
1
1
LBMA (S) = inf
(L?? (S) + ||?? ||2 ) + log(1 + T )
??
2
2
(1)
(T )
Let Y = {y , . . . , y } and X = {1, . . . , 1}. By the chain rule of conditional probabilities, LBMA (S) = ? log p(Y |X) (where p is the Gaussian linear regression model), and
PT
q?s loss is t=1 `q (t) = ? log q(Y |X). For any predictive distribution q that differs from
p, there must exist some sequence S such that ? log q(Y |X) is greater than ? log p(Y |X)
(since probabilities are normalized). Such a sequence proves the result for n = 1.
The modification for n dimensions follows: S is broken into bT /nc subsequences where
(t)
in every subsequence only one dimension has xk = 1 (and the other dimensions are set to
0). The result follows due to the additivity of the losses on these subsequences.
3
MAP Estimation
We now present bounds for MAP algorithms for both Gaussian linear regression (i.e., ridge
regression) and logistic regression. These algorithms use the maximum ??t?1 of pt?1 (?) to
form their predictive distribution p(y|x(t) , ??t?1 ) at time t, as opposed to BMA?s predictive
distribution of p(y|x(t) , St?1 ). As expected these bounds are weaker than BMA, though
perhaps not unreasonably so.
3.1 The Square Loss and Ridge Regression
Before we provide the MAP bound, let us first present the form of the posteriors and the
Pt
T
predictions for Gaussian linear regression. Define At = ?12 In + ?12 i=1 x(i) x(i) , and
Pt
bt = i=1 x(i) y (i) . We now have that
?t ,
pt (?) = p(?|St ) = N ?; ??t , ?
(6)
?1
?
where ??t = A?1
t bt , and ?t = At . Also, the predictions at time t + 1 are given by
p(y (t+1) |x(t+1) , St ) = N y (t+1) ; y?t+1 , s2t+1
(7)
T
? t x(t+1) + ? 2 . In contrast, the prediction of a
where y?t+1 = ??tT x(t+1) , s2t+1 = x(t+1) ?
?
fixed expert using parameter ? would be
p(y (t) |x(t) , ?? ) = N y (t) ; yt? , ? 2 ,
(8)
where yt? = ?? T x(t) .
Now the BMA loss is:
LBMA (S) =
T
q
X
1 (t) ?T (t) 2
2?s2t
(y
?
?
x
)
+
log
t?1
2
2s
t
t=1
(9)
Importantly, note how Bayes is adaptively weighting the squared term with the inverse
variances 1/st (which depend on the current observation x(t) ). The logloss of using a fixed
expert ?? is just:
L?? (S) =
T
X
?
1
(y (t) ? ?? T x(t) )2 + log 2?? 2
2
2?
t=1
(10)
The MAP procedure (referred to as ridge regression) uses p(y|x(t) , ??t?1 ) which has a fixed
variance. Hence, the MAP loss is essentially the square loss and we define it as such:
T
T
X
X
T
e ?? (S) = 1
e MAP (S) = 1
(y (t) ? ??t?1
x(t) )2 , L
(y (t) ? ?? T x(t) )2 ,
L
2 t=1
2 t=1
(11)
where ??t is the MAP estimate (see Equation 6).
Corollary 3.1: Let ? 2 = ? 2 + ? 2 . For all S such that ||x(t) || ? 1 and for all ?? , we have
2
2
2
2
e MAP (S) ? ? L
e ?? (S) + ? ||?? ||2 + ? n log 1 + T ?
L
?2
2? 2
2
?2 n
Proof: Using Equations (9,10) and Theorem 2.2, we have
T
X
1 (t) ?T (t) 2
? ?t?1 x )
2 (y
2s
t
t=1
?
T
X
1
1
(y (t) ? ?? T x(t) )2 + 2 ||?? ||2
2
2?
2?
t=1
?
X
T
T c? 2
2?? 2
n
+ log 1 +
+
log p
2
n
2?s2t
t=1
Equations (6, 7) imply that ? 2 ? s2t ? ? 2 + ? 2 . Using this, the result follows by noting
that the last term is negative and by multiplying both sides of the equation by ? 2 + ? 2 .
We might have hoped that MAP were more competitive in that the leading coefficient,
e ?? (S) term in the bound, be 1 (similar to Theorem 2.2) rather than ? 22 .
in front of the L
?
Crudely, the reason that MAP is not as effective as BMA is that MAP does not take into
account the uncertainty in its predictions?thus the squared terms cannot be reweighted to
take variance into account (compare Equations 9 and 11).
Some previous (non-Bayesian) algorithms did in fact have bounds with this coefficient
being unity. Vovk (2001) provides such an algorithm, though this algorithm differs from
MAP in that its predictions at time t are a nonlinear function of x(t) (it uses At instead
of At?1 at time t). Foster (1991) provides a bound with this coefficient being 1 with
more restrictive assumptions. Azoury and Warmuth (2001) also provide a bound with a
coefficient of 1 by using a MAP procedure with ?clipping.? (Their algorithm thresholds the
T
prediction y?t = ??t?1
x(t) if it is larger than some upper bound. Note that we do not assume
any upper bound on y (t) .)
As the following lower bound shows, it is not possible for the MAP linear regression algoe ?? (S), with a reasonable regret bound. A similar lower
rithm to have a coefficient of 1 for L
bound is in Vovk (2001), which doesn?t apply to our setting where we have the additional
constraint ||x(t) || ? 1.
Theorem 3.2: Let ? 2 = ? 2 + ? 2 . There exists a sequence S with ||x(t) || ? 1 such that
e MAP (S) ? inf (L
e ?? (S) + 1 ||?? ||2 ) + ?(T )
L
??
2
Proof: (sketch)
Let S be a length T + 1 sequence, with n = 1, where for the first T steps,
?
x(t) = 1/ T and y (t) = 1, and at T + 1, x(T +1) = 1 and y (T +1) = 0. Here, one can show
e ?? (S) + 1 ||?? ||2 ) = T /4 and L
e MAP (S) ? 3T /8, and the result follows.
that inf ?? (L
2
3.2 Logistic Regression
MAP estimation is often used for regularized logistic regression, since it requires only
solving a convex program (while BMA has to deal with a high dimensional integral over
? that is intractable to compute exactly). Letting ??t?1 be the maximum of the posterior
PT
pt?1 (?), define LMAP (S) = t=1 ? log p(y (t) |x(t) , ??t?1 ). As with the square loss case,
the bound we present is multiplicatively worse (by a factor of 4).
Theorem 3.3: In the logistic regression model with ? ? 0.5, we have that for all sequences
S such that ||x(t) || ? 1 and y (t) ? {0, 1} and for all ??
T ?2
2
LMAP (S) ? 4L?? (S) + 2 ||?? ||2 + 2n log 1 +
?
n
Proof: (sketch) Assume n = 1 (the general case is analogous). The proof consists of
showing that `??t?1 (t) = ? log p(y (t) |x(t) , ??t?1 ) ? 4`BMA (t). Without loss of generality,
assume y (t) = 1 and xR(t) ? 0, and for convenience, we just write x instead of x(t) . Now
the BMA prediction is ? p(1|?, x)pt?1 (?)d?, and `BMA (t) is the negative log of this. Note
? = ? gives probability 1 for y (t) = 1 (and this setting of ? minimizes the loss at time t).
Since we do not have a closed form solution of the posterior pt?1 , let us work with another
distribution q(?) in lieu of pt?1 (?) that satisfies certain properties. Define pq =
R
p(1|?,
x)q(?)d?, which can be viewed as the prediction using q rather than the posterior.
?
We choose q to be the rectification of the Gaussian N (?; ??t?1 , ? 2 In ), such that there is
positive probability only for ? ? ??t?1 (and the distribution is renormalized). With this
choice, we first show that the loss of q, ? log pq , is less than or equal to `BMA (t). Then we
complete the proof by showing that `??t?1 (t) ? ?4 log pq , since ? log pq ? `BMA (t).
Consider the q which maximizes pq subject to the following constraints: let q(?) have its
maximum at ??t?1 ; let q(?) = 0 if ? < ??t?1 (intuitively, mass to the left of ??t?1 is just
making the pq smaller); and impose the constraint that ?(log q(?))00 ? 1/? 2 . We now
argue that for such a q, ? log pq ? `BMA (t). First note that due to the Gaussian prior p0 , it
is straightforward to show that ?(log pt?1 )00 (?) ? ?12 (the prior imposes some minimum
curvature). Now if this posterior pt?1 were rectified (with support only for ? ? ??t?1 )
and renormalized, then such a modified distribution clearly satisfies the aforementioned
constraints, and it has loss less than the loss of pt?1 itself (since the rectification only
increases the prediction). Hence, the maximizer, q, of pq subject to the constraints has loss
less than that of pt?1 , i.e. ? log pq ? `BMA (t).
We now show that such a maximal q is the (renormalized) rectification of the Gaussian
N (?; ??t?1 , ? 2 In ), such that there is positive probability only for ? > ??t?1 . Assume some
other q2 satisfied these constraints and maximized pq . It cannot be that q2 (??t?1 ) < q(??t?1 ),
else one can show q2 would not be normalized (since with q2 (??t?1 ) < q(??t?1 ), the curvature constraint imposes that this q2 cannot cross q). It also cannot be that q2 (??t?1 ) >
q(??t?1 ). To see this, note that normalization and curvature imply that q2 must cross pt only
once. Now a sufficiently slight perturbation of this crossing point to the left, by shifting
more mass from the left to the right side of the crossing point, would not violate the curvature constraint and would result in a new distribution with larger pq , contradicting the
maximality of q2 . Hence, we have that q2 (??t?1 ) = q(??t?1 ). This, along with the curvature
constraint and normalization, imply that the rectified Gaussian, q, is the unique solution.
To complete the proof, we show ` ? (t) = ? log p(1|x, ??t?1 ) ? ?4 log pq . We consider
?t?1
two cases, ??t?1 < 0 and ??t?1 ? 0. We start with the case ??t?1 < 0. Using the boundedness
of the derivative |? log p(1|x, ?)/??| < 1 and that q only has support for ? ? ??t?1 , we have
Z
pq =
exp(log p(1|x, ?))q(?)d?
Z?
?
exp log(p(1|x, ??t?1 ) + ? ? ??t?1 q(?)d? ? 1.6p(1|x, ??t?1 )
?
R
where we have used that ? exp(? ? ??t?1 )q(?)d? < 1.6 (which can be verified numerically
using the definition of q with ? ? 0.5). Now observe that for ??t?1 ? 0, we have the lower
bound ? log p(1|x, ??t?1 ) ? log 2. Hence, ? log pq ? ? log p(1|x, ??t?1 ) ? log 1.6 ?
(? log p(1|x, ??t?1 ))(1 ? log 1.6/ log 2) ? 0.3`??t?1 (t), which shows `??t?1 (t) ? ?4 log pq .
?
Now for
R the case ?t?1 ? 0. Let ? be the sigmoid function, so p(1|x, ?) = ?(?x) and
pq = ? ?(x?)q(?)d?. Since the sigmoid is concave
R for ? >0 and, for this case, q only has
support from positive ?, we have that pq ? ? x ? ?q(?)d? . Using the definition of q, we
then have that pq ? ?(x(??t?1 + ?)) ? ?(??t?1 + ?), where the last inequality follows from
??t?1 + ? > 0 and x ? 1. Using properties of ?, one can show |(log ?)0 (z)| < ? log ?(z)
(for all z). Hence, for all ? ? ??t?1 , |(log ?)0 (?)| < ? log ?(?) ? ? log ?(??t?1 ). Using
this derivative condition along with the previous bound on pq , we have that ? log pq ?
? log ?(??t?1 + ?) ? (? log ?(??t?1 ))(1 ? ?) = `??t?1 (t)(1 ? ?), which shows that
` ? (t) ? ?4 log pq (since ? ? 0.5). This proves the claim when ??t?1 ? 0.
?t?1
Acknowledgments. We thank Dean Foster for numerous helpful discussions. This work
was supported by the Department of the Interior/DARPA under contract NBCHD030010.
References
Azoury, K. S. and Warmuth, M. (2001). Relative loss bounds for on-line density estimation with the
exponential family of distributions. Machine Learning, 43(3).
Cesa-Bianchi, N., Freund, Y., Haussler, D., Helmbold, D., Schapire, R., and Warmuth, M. (1997).
How to use expert advice. J. ACM, 44.
Cesa-Bianchi, N., Helmbold, D., and Panizza, S. (1998). On Bayes methods for on-line boolean
prediction. Algorithmica, 22.
Dawid, A. (1984). Statistical theory: The prequential approach. J. Royal Statistical Society.
Foster, D. P. (1991). Prediction in the worst case. Annals of Statistics, 19.
Freund, Y. and Schapire, R. (1999). Adaptive game playing using multiplicative weights. Games and
Economic Behavior, 29:79?103.
Freund, Y., Schapire, R., Singer, Y., and Warmuth, M. (1997). Using and combining predictors that
specialize. In STOC.
Grunwald, P. (2005). A tutorial introduction to the minimum description length principle.
McCullagh, P. and Nelder, J. A. (1989). Generalized Linear Models (2nd ed.). Chapman and Hall.
Ng, A. Y. and Jordan, M. (2001). Convergence rates of the voting Gibbs classifier, with application
to Bayesian feature selection. In Proceedings of the 18th Int?l Conference on Machine Learning.
Vovk, V. (2001). Competitive on-line statistics. International Statistical Review, 69.
| 2637 |@word norm:1 nd:1 covariance:2 p0:7 simplifying:1 boundedness:1 current:1 comparing:2 written:1 must:2 warmuth:8 xk:1 realism:1 completeness:1 provides:4 simpler:1 along:2 s2t:5 prove:2 consists:1 specialize:1 expected:3 indeed:1 behavior:1 examine:1 provided:2 maximizes:1 mass:2 interpreted:1 minimizes:1 q2:9 every:1 voting:1 concave:1 exactly:1 classifier:1 before:1 positive:3 might:1 studied:1 lmap:2 unique:1 acknowledgment:1 practice:1 regret:1 differs:4 xr:1 procedure:4 get:1 cannot:4 convenience:1 interior:1 selection:1 map:22 dean:1 yt:2 straightforward:1 convex:2 helmbold:2 rule:4 haussler:1 importantly:3 deriving:1 analogous:1 annals:1 pt:20 suppose:1 us:4 dawid:2 crossing:2 expensive:1 particularly:1 worst:6 prequential:2 broken:1 renormalized:3 depend:1 solving:2 panizza:1 predictive:5 learner:1 darpa:1 additivity:1 fast:1 effective:1 fy00:6 choosing:1 stanford:1 widely:3 larger:2 statistic:3 itself:1 online:12 sequence:11 differentiable:1 maximal:1 combining:1 description:1 convergence:1 generating:1 andrew:1 progress:1 predicted:1 come:2 implies:1 correct:1 require:1 preliminary:1 hold:2 sufficiently:1 considered:1 hall:1 normal:1 exp:7 claim:1 estimation:4 bma:15 label:1 clearly:1 gaussian:19 modified:1 rather:2 corollary:1 contrast:2 adversarial:4 sense:1 posteriori:2 helpful:1 typically:1 bt:3 going:1 interested:1 aforementioned:1 equal:1 once:1 ng:3 chapman:1 identical:2 adversarially:1 algorithmica:1 consisting:1 mdl:1 truly:1 analyzed:1 chain:2 logloss:9 bregman:1 integral:1 continuing:1 taylor:1 theoretical:1 minimal:1 modeling:2 boolean:2 clipping:2 tractability:1 cost:1 predictor:1 too:1 front:1 adaptively:1 st:13 density:2 international:1 probabilistic:1 contract:1 together:2 continuously:1 squared:2 cesa:4 espoused:1 opposed:2 choose:1 possibly:1 satisfied:1 worse:1 expert:7 derivative:2 leading:1 yp:1 account:2 coefficient:5 int:1 depends:1 multiplicative:2 later:1 closed:1 competitive:8 bayes:5 start:2 square:7 variance:4 who:1 maximized:1 bayesian:23 multiplying:1 rectified:2 suffers:1 ed:1 definition:3 against:2 proof:10 recall:1 knowledge:1 focusing:1 though:8 generality:2 furthermore:1 governing:1 just:4 crudely:1 retrospect:1 sketch:3 nonlinear:1 maximizer:1 logistic:11 perhaps:1 normalized:2 true:1 equality:1 hence:5 deal:2 reweighted:1 round:1 game:2 generalized:5 ridge:5 tt:1 complete:2 variational:2 sigmoid:2 discussed:2 fare:1 slight:1 theirs:1 numerically:1 significant:1 gibbs:1 had:1 pq:21 curvature:5 posterior:7 inf:4 reverse:1 certain:2 inequality:1 seen:1 minimum:2 greater:1 somewhat:1 floor:1 additional:1 impose:1 full:2 violate:1 sham:1 believed:1 cross:2 devised:2 prediction:18 variant:3 regression:37 essentially:1 expectation:3 iteration:1 sometimes:2 normalization:2 sleeping:1 else:1 grow:1 extra:1 subject:2 jordan:2 noting:1 pennsylvania:1 economic:1 penalty:2 suffer:1 useful:2 involve:1 tune:1 schapire:4 exist:1 tutorial:1 write:1 putting:1 threshold:1 verified:1 sum:1 inverse:1 uncertainty:1 family:2 throughout:1 almost:1 reasonable:1 comparable:2 bound:39 constraint:9 argument:1 department:3 cleverly:1 smaller:1 slightly:1 unity:1 kakade:1 modification:1 making:1 intuitively:1 restricted:1 computationally:1 equation:8 rectification:3 previously:1 singer:1 letting:2 lieu:1 tightest:1 multiplied:1 apply:2 observe:1 appropriate:1 shortly:1 unreasonably:2 denotes:3 instant:1 restrictive:1 prof:3 society:1 ingenious:1 usual:1 distance:1 thank:1 argue:1 fy:14 reason:1 length:2 multiplicatively:1 balance:1 nc:1 stoc:1 relate:1 favorably:1 negative:2 perform:2 bianchi:4 upper:4 observation:1 finite:1 rn:3 perturbation:1 arbitrary:4 namely:1 kl:6 distinction:1 nbchd030010:1 adversary:2 program:1 including:1 royal:1 shifting:1 regularized:3 scheme:1 maximality:1 imply:3 numerous:1 started:1 prior:8 literature:4 l2:1 review:1 relative:3 fy0:2 freund:6 loss:27 generation:1 var:2 incurred:2 imposes:2 foster:6 principle:1 playing:1 supported:1 last:4 transpose:1 side:3 weaker:1 taking:1 dimension:3 cumulative:1 doesn:1 made:1 adaptive:1 assumed:2 nelder:2 subsequence:4 decade:1 expansion:1 complex:1 did:1 main:1 azoury:6 motivation:1 n2:1 contradicting:1 advice:1 referred:1 grunwald:2 rithm:1 wish:1 lq:5 exponential:1 rederive:1 weighting:1 theorem:9 xt:1 showing:3 exists:2 intractable:1 hoped:1 entropy:1 tc:2 logarithmic:1 ez:2 satisfies:2 acm:1 conditional:2 goal:3 identity:1 viewed:1 mccullagh:2 specifically:1 vovk:8 averaging:5 lemma:5 support:3 violated:1 philosophy:1 |
1,803 | 2,638 | Synergistic Face Detection and Pose Estimation
with Energy-Based Models
Margarita Osadchy
NEC Labs America
Princeton NJ 08540
[email protected]
Matthew L. Miller
NEC Labs America
Princeton NJ 08540
[email protected]
Yann Le Cun
The Courant Institute
New York University
[email protected]
Abstract
We describe a novel method for real-time, simultaneous multi-view face
detection and facial pose estimation. The method employs a convolutional network to map face images to points on a manifold, parametrized
by pose, and non-face images to points far from that manifold. This
network is trained by optimizing a loss function of three variables: image, pose, and face/non-face label. We test the resulting system, in a
single configuration, on three standard data sets ? one for frontal pose,
one for rotated faces, and one for profiles ? and find that its performance
on each set is comparable to previous multi-view face detectors that can
only handle one form of pose variation. We also show experimentally
that the system?s accuracy on both face detection and pose estimation is
improved by training for the two tasks together.
1 Introduction
The detection of human faces in natural images and videos is a key component in a wide
variety of applications of human-computer interaction, search and indexing, security, and
surveillance. Many real-world applications would profit from multi-view detectors that can
detect faces under a wide range of poses: looking left or right (yaw axis), up or down (pitch
axis), or tilting left or right (roll axis).
In this paper we describe a novel method that not only detects faces independently of their
poses, but simultaneously estimates those poses. The system is highly-reliable, runs at near
real time (5 frames per second on standard hardware), and is robust against variations in
yaw (?90? ), roll (?45? ), and pitch (?60? ).
The method is motivated by the idea that multi-view face detection and pose estimation are
so closely related that they should not be performed separately. The tasks are related in the
sense that they must be robust against the same sorts of variation: skin color, glasses, facial
hair, lighting, scale, expressions, etc. We suspect that, when trained together, each task can
serve as an inductive bias for the other, yielding better generalization or requiring fewer
training examples [2].
To exploit the synergy between these two tasks, we train a convolutional network to map
face images to points on a face manifold, and non-face images to points far away from
that manifold. The manifold is parameterized by facial pose. Conceptually, we can view
the pose parameter as a latent variable that can be inferred through an energy-minimization
process [4]. To train the machine we derive a new type of discriminative loss function that
is tailored to such detection tasks.
Previous Work: Learning-based approaches to face detection abound, including real-time
methods [16], and approaches based on convolutional networks [15, 3]. Most multi-view
systems take a view-based approach, which involves building separate detectors for different views and either applying them in parallel [10, 14, 13, 7] or using a pose estimator to
select a detector [5]. Another approach is to estimate and correct in-plane rotations before
applying a single pose-specific detector [12]. Closer to our approach is that of [8], in which
a number of Support Vector Regressors are trained to approximate smooth functions, each
of which has a maximum for a face at a particular pose. Another machine is trained to convert the resulting values to estimates of poses, and a third is trained to convert the values
into a face/non-face score. The resulting system is very slow.
2 Integrating face detection and pose estimation
To exploit the posited synergy between face detection and pose estimation, we must design
a system that integrates the solutions to the two problems. We hope to obtain better results
on both tasks, so this should not be a mere cascaded system in which the answer to one
problem is used to assist in solving the other. Both answers must be derived from one
underlying analysis of the input, and both tasks must be trained together.
Our approach is to build a trainable system that can map raw images X to points in a
low-dimensional space. In that space, we pre-define a face manifold F (Z) that we parameterize by the pose Z. We train the system to map face images with known poses to the
corresponding points on the manifold. We also train it to map non-face images to points far
away from the manifold. Proximity to the manifold then tells us whether or not an image
is a face, and projection to the manifold yields an estimate of the pose.
Parameterizing the Face Manifold: We will now describe the details of the parameterizations of the face manifold. Let?s start with the simplest case of one pose parameter
Z = ?, representing, say, yaw. If we want to preserve the natural topology and geometry
of the problem, the face manifold under yaw variations in the interval [?90? , 90? ] should
be a half circle (with constant curvature). We embed this half-circle in a three-dimensional
space using three equally-spaced shifted cosines.
? ?
(1)
Fi (?) = cos(? ? ?i ); i = 1, 2, 3; ? = [? , ]
2 2
When we run the network on an image X, it outputs a vector G(X) with three components
that can be decoded analytically into corresponding pose angle:
P3
Gi (X) cos(?i )
? = arctan Pi=1
(2)
3
i=1 Gi (X) sin(?i )
The point on the manifold closest to G(X) is just F (?).
The same idea can be applied to any number of pose parameters. Let us consider the set
of all faces with yaw in [?90, 90] and roll in [?45, 45]. In an abstract way, this set is
isomorphic to a portion of the surface of a sphere. Consequently, we encode the pose with
the product of the cosines of the two angles:
Fij (?, ?) = cos(? ? ?i ) cos(? ? ?j ); i, j = 1, 2, 3;
(3)
For convenience we rescale the roll angles to the range of [?90, 90]. With these parameterizations, the manifold has constant curvature, which ensures that the effect of errors will
be the same regardless of pose. Given nine components of the network?s output Gij (X),
we compute the corresponding pose angles as follows:
P
P
cc = Pij Gij (X) cos(?i ) cos(?j ); cs = Pij Gij (X) cos(?i ) sin(?j )
sc =
ss =
ij Gij (X) sin(?i ) cos(?j );
ij Gij (X) sin(?i ) sin(?j )
(4)
? = 0.5(atan2(cs + sc, cc ? ss) + atan2(sc ? cs, cc + ss))
? = 0.5(atan2(cs + sc, cc ? ss) ? atan2(sc ? cs, cc + ss))
Note that the dimension of the face manifold is much lower than that of the embedding
space. This gives ample space to represent non-faces away from the manifold.
3 Learning Machine
To build a learning machine for the proposed approach we refer to the Minimum Energy
Machine framework described in [4].
Energy Minimization Framework: We can view our system as a scalar-value function
EW (Y, Z, X), where X and Z are as defined above, Y is a binary label (Y = 1 for face,
Y = 0 for a non-face), and W is a parameter vector subject to learning. EW (Y, Z, X)
can be interpreted as an energy function that measures the degree of compatibility between
X, Z, Y . If X is a face with pose Z, then we want: EW (1, Z, X) ? EW (0, Z 0 , X) for
any pose Z 0 , and EW (1, Z 0 , X) ? EW (1, Z, X) for any pose Z 0 6= Z.
Operating the machine consists in clamping X to the observed value (the image), and
finding the values of Z and Y that minimize EW (Y, Z, X):
(Y , Z) = argminY ?{Y }, Z?{Z} EW (Y, Z, X)
(5)
where {Y } = {0, 1} and {Z} = [?90, 90]?[?45, 45] for yaw and roll variables. Although
this inference process can be viewed probabilistically as finding the most likely configuration of Y and Z according to a model that attributes high probabilities to low-energy
configurations (e.g. a Gibbs distribution), we view it as a non probabilistic decision making process. In other words, we make no assumption as to the finiteness of integrals over
{Y }and {Z}that would be necessary for a properly normalized probabilistic model. This
gives us considerable flexibility in the choice of the internal architecture of EW (Y, Z, X).
Our energy function for a face EW (1, Z, X) is defined as the distance between the point
produced by the network GW (X) and the point with pose Z on the manifold F (Z):
EW (1, Z, X) = kGW (X) ? F (Z)k
(6)
The energy function for a non-face EW (0, Z, X) is equal to a constant T that we can
interpret as a threshold (it is independent of Z and X). The complete energy function is:
EW (Y, Z, X) = Y kGW (X) ? F (Z)k + (1 ? Y )T
(7)
The architecture of the machine is depicted in Figure 1. Operating this machine (finding the output label and pose with the smallest energy) comes down to first finding:
Z = argminZ?{Z} ||GW (X) ? F (Z)||, and then comparing this minimum distance,
kGW (X) ? F (Z)k, to the threshold T . If it smaller than T , then X is classified as a face,
otherwise X is classified as a non-face. This decision is implemented in the architecture as
a switch, that depends upon the binary variable Y .
Convolutional Network: We employ a Convolutional Network as the basic architecture for
the GW (X) image-to-face-space mapping function. Convolutional networks [6] are ?endto-end? trainable system that can operate on raw pixel images and learn low-level features
and high-level representation in an integrated fashion. Convolutional nets are advantageous
because they easily learn the types of shift-invariant local features that are relevant to image
recognition; and more importantly, they can be replicated over large images (swept over
every location) at a fraction of the cost of replicating more traditional classifiers [6]. This
is a considerable advantage for building real-time systems.
We employ a network architecture similar to LeNet5 [6]. The difference is in the number
of maps. In our architecture we have 8 feature maps in the bottom convolutional and
subsampling layers and 20 maps in the next two layers. The last layer has 9 outputs to
encode two pose parameters.
Training with a Discriminative Loss Function for Detection: We define the loss function
as follows:
1 X
1 X
L1 (W, Z i , X i ) +
L0 (W, X i )
(8)
L(W ) =
|S1 |
|S0 |
i?S1
i?S0
Figure 1: Architecture of the Minimum Energy Machine.
where S1 is the set of training faces, S0 the set of non-faces, L1 (W, Z i , X i ) and L0 (W, X i )
are loss functions for a face sample (with a known pose) and non-face, respectively1 .
The loss L(W ) should be designed so that its minimization for a particular positive training
sample (X i , Z i , 1), will make EW (1, Z i , X i ) < EW (Y, Z, X i ) for Y 6= Y i or Z 6= Z i .
To satisfy this, it is sufficient to make EW (1, Z i , X i ) < EW (0, Z, X i ). For a particular negative training sample (X i , 0), minimizing the loss should make EW (1, Z, X i ) >
EW (0, Z, X i ) = T for any Z. To satisfy this, it is sufficient to make EW (1, Z, X i ) > T .
Let W be the current parameter value, and W 0 be the parameter value after an update
caused by a single sample. To cause the machine to achieve the desired behavior, we need
the parameter update to decrease the difference between the energy of the desired label and
the energy of the undesired label. In our case, since EW (0, Z, X) = T is constant, the
following condition on the update is sufficient to ensure the desired behavior:
Condition 1. for a face example (X, Z, 1), we must have: EW 0 (1, Z, X) < EW (1, Z, X)
For a non-face example (X, 1), we must have: EW 0 (1, Z, X) > EW (1, Z, X)
We choose the following forms for L1 and L0 :
L1 (W, 1, Z, X) = EW (1, Z, X)2 ;
L0 (W, 0, X) = K exp[?E(1, Z, X)]
(9)
where K is a positive constant.
Next we show that minimizing (9) with an incremental gradient-based algorithm will satisfy
condition 1. With gradient-based optimization algorithms, the parameter update formula
?L
. where A is a judiciously chosen symmetric
is of the form: ?W = W 0 ? W = ??A ?W
positive semi-definite matrix, and ? is a small positive constant.
2
(1,Z,X)
For Y = 1 (face): An update step will change the parameter by ?W = ??A ?EW ?W
=
(1,Z,X)
?2?EW (1, Z, X)A ?EW?W
. To first order (for small values of ?), the resulting change
in EW (1, Z, X) is given by:
?EW (1, Z, X)
?W
T
?W = ?2?EW (1, Z, X)
?EW (1, Z, X)
?W
T
A
?EW (1, Z, X)
<0
?W
because EW (1, Z, X) > 0 (it?s a distance), and the quadratic form is positive. Therefore
EW 0 (1, Z, X) < EW (1, Z, X).
1
Although face samples whose pose is unknown can easily be accommodated, we will not discuss
this possibility here.
95
Percentage of yaws correctly estimated
100
95
Percentage of faces detected
100
90
85
80
75
70
65
60
Pose + detection
Detection only
55
50
0
2
4
6
8
10
12
14
16
18
20
90
85
80
75
70
65
60
Pose + detection
Pose only
55
50
False positive rate
0
5
10
15
20
25
30
Yaw-error tolerance (degrees)
Figure 2: Synergy test. Left: ROC curves for the pose-plus-detection and detection-only networks.
Right: frequency with which the pose-plus-detection and pose-only networks correctly estimated the
yaws within various error tolerances.
For Y = 0 (non-face): An update step will change the parameter by ?W =
(1,Z,X)
??A ?K exp[?E(1,Z,X)]
= ?K exp[?EW (1, Z, X)] ?EW?W
. To first order (for small
?W
values of ?), the resulting change in EW (1, Z, X) is given by:
?EW (1, Z, X)
?W
T
?W = ?K exp[?EW (1, Z, X)]
?EW (1, Z, X)
?W
T
A
?EW (1, Z, X)
>0
?W
Therefore EW 0 (1, Z, X) > EW (1, Z, X).
Running the Machine: Our detection system works on grayscale images
? and it applies
the network to each image at a range of scales, stepping by a factor of 2. The network
is replicated over the image at each scale, stepping by 4 pixels in x and y (this step size
is a consequence of having two, 2x2 subsampling layers). At each scale and location,
the network outputs are compared to the closest point on the manifold, and the system
collects a list of all instances closer than our detection threshold. Finally, after examining
all scales, the system identifies groups of overlapping detections in the list and discards all
but the strongest (closest to the manifold) from each group. No attempt is made to combine
detections or apply any voting scheme. We have implemented the system in C. The system
can detect, locate, and estimate the pose of faces that are between 40 and 250 pixels high
in a 640 ? 480 image at roughly 5 frames per second on a 2.4GHz Pentium 4.
4 Experiments and results
Using the above architecture, we built a detector to locate faces and estimate two pose
parameters: yaw from left to right profile, and in-plane rotation from ?45 to 45 degrees.
The machine was trained to be robust against pitch variation.
In this section, we first describe the training regimen for this network, and then give the
results of two sets of experiments. The first set of experiments tests whether training for
the two tasks together improves performance on both. The second set allows comparisons
between our system and other published multi-view detectors.
Training: Our training set consisted of 52, 850, 32x32-pixel faces from natural images
collected at NEC Labs and hand annotated with appropriate facial poses (see [9] for a
description of how the annotation was done). These faces were selected from a much
larger annotated set to yield a roughly uniform distribution of poses from left profile to right
profile, with as much variation in pitch as we could obtain. Our initial negative training data
consisted of 52, 850 image patches chosen randomly from non-face areas of a variety of
images. For our second set of tests, we replaced half of these with image patches obtained
by running the initial version of the detector on our training images and collecting false
detections. Each training image was used 5 times during training, with random variations
95
Percentage of poses correctly estimated
100
95
Percentage of faces detected
100
90
85
80
75
70
65
Frontal
Rotated in plane
Profile
60
55
50
0
0.5
1
1.5
2
2.5
3
3.5
False positives per image
4
4.5
5
90
85
80
75
70
65
60
In-plane rotation
Yaw
55
50
0
5
10
15
20
25
30
Pose-error tolerance (degrees)
Figure 3: Results on standard data sets. Left: ROC curves for our detector on the three data
sets. The x axis is the average number of false positives per image over all three sets, so each point
corresponds to a single detection threshold. Right: frequency with which yaw and roll are estimated
within various error tolerances.
?
?
in scale (from x 2 to x(1 + 2)), in-plane rotation (?45? ), brightness (?20), contrast
(from 0.8 to 1.3).
To train the network, we made 9 passes through this data, though it mostly converged after
about the first 6 passes. Training was performed using LUSH [1], and the total training
time was about 26 hours on a 2Ghz Pentium 4. At the end of training, the network had
converged to an equal error rate of 5% on the training data and 6% on a separate test set of
90,000 images.
Synergy tests: The goal of the synergy test was to verify that both face detection and pose
estimation benefit from learning and running in parallel. To test this claim we built three
networks with almost identical architectures, but trained to perform different tasks. The
first one was trained for simultaneous face detection and pose estimation (combined), the
second was trained for detection only and the third for pose estimation only. The ?detection
only? network had only one output for indicating whether or not its input was a face. The
?pose only? network was identical to the combined network, but trained on faces only (no
negative examples). Figure 2 shows the results of running these networks on our 10,000
test images. In both these graphs, we see that the pose-plus-detection network had better
performance, confirming that training for each task benefits the other.
Standard data sets: There is no standard data set that tests all the poses our system is
designed to detect. There are, however, data sets that have been used to test more restricted
face detectors, each set focusing on a particular variation in pose. By testing a single
detector with all of these sets, we can compare our performance against published systems.
As far as we know, we are the first to publish results for a single detector on all these data
sets. The details of these sets are described below:
? MIT+CMU [14, 11] ? 130 images for testing frontal face detectors. We count 517 faces
in this set, but the standard tests only use a subset of 507 faces, because 10 faces are in
the wrong pose or otherwise not suitable for the test. (Note: about 2% of the faces in the
standard subset are badly-drawn cartoons, which we do not intend our system to detect.
Nevertheless, we include them in the results we report.)
? TILTED [12] ? 50 images of frontal faces with in-plane rotations. 223 faces out of 225
are in the standard subset. (Note: about 20% of the faces in the standard subset are outside
of the ?45? rotation range for which our system is designed. Again, we still include these
in our results.)
? PROFILE [13] ? 208 images of faces in profile. There seems to be some disagreement
about the number of faces in the standard set of annotations: [13] reports using 347 faces
of the 462 that we found, [5] reports using 355, and we found 353 annotations. However,
these discrepencies should not significantly effect the reported results.
We counted a face as being detected if 1) at least one detection lay within a circle centered
on the midpoint between the eyes, with a radius equal to 1.25 times the distance from that
point to the midpoint of the mouth, and 2) that detection came at a scale within a factor of
Figure 4: Some example face detections. Each white box shows the location of a detected face. The
angle of each box indicates the estimated in-plane rotation. The black crosshairs within each box
indicate the estimated yaw.
Data set ?
False positives per image ?
Our detector
Jones & Viola [5] (tilted)
Jones & Viola [5] (profile)
Rowley et al [11]
Schneiderman & Kanade [13]
TILTED
4.42
90%
90%
26.90
97%
95%
x
89%
PROFILE
.47
67%
x
70%
83%
x
96%
x
3.36
83%
86%
93%
MIT+CMU
.50
83%
1.28
88%
x
x
x
x
Table 1: Comparisons of our results with other multi-view detectors. Each column shows the detection rates for a given average number of false positives per image (these rates correspond to those for
which other authors have reported results). Results for real-time detectors are shown in bold. Note
that ours is the only single detector that can be tested on all data sets simultaneously.
two of the correct scale for the face?s size. We counted a detection as a false positive if it
did not lie within this range for any of the faces in the image, including those faces not in
the standard subset.
The left graph in Figure 3 shows ROC curves for our detector on the three data sets. Figure 4
shows a few results on various poses. Table 1 shows our detection rates compared against
other systems for which results were given on these data sets. The table shows that our
results on the TILTED and PROFILE sets are similar to those of the two Jones & Viola
detectors, and even approach those of the Rowley et al and Schneiderman & Kanade nonreal-time detectors. Those detectors, however, are not designed to handle all variations in
pose, and do not yield pose estimates.
The right side of Figure 3 shows our performance at pose estimation. To make this graph,
we fixed the detection threshold at a value that resulted in about 0.5 false positives per
image over all three data sets. We then compared the pose estimates for all detected faces
(including those not in the standard subsets) against our manual pose annotations. Note that
this test is more difficult than typical tests of pose estimation systems, where faces are first
localized by hand. When we hand-localize these faces, 89% of yaws and 100% of in-plane
rotations are correctly estimated to within 15? .
5 Conclusion
The system we have presented here integrates detection and pose estimation by training
a convolutional network to map faces to points on a manifold, parameterized by pose,
and non-faces to points far from the manifold. The network is trained by optimizing a loss
function of three variables ? image, pose, and face/non-face label. When the three variables
match, the energy function is trained to have a small value, when they do not match, it is
trained to have a large value.
This system has several desirable properties:
? The use of a convolutional network makes it fast. At typical webcam resolutions, it can
process 5 frames per second on a 2.4Ghz Pentium 4.
? It is robust to a wide range of poses, including variations in yaw up to ?90? , in-plane
rotation up to ?45? , and pitch up to ?60? . This has been verified with tests on three
standard data sets, each designed to test robustness against a single dimension of pose
variation.
? At the same time that it detects faces, it produces estimates of their pose. On the standard
data sets, the estimates of yaw and in-plane rotation are within 15? of manual estimates
over 80% and 95% of the time, respectively.
We have shown experimentally that our system?s accuracy at both pose estimation and face
detection is increased by training for the two tasks together.
References
[1] L. Bottou and Y. LeCun. The Lush Manual. http://lush.sf.net, 2002.
[2] R. Caruana. Multitask learning. Machine Learning, 28:41?75, 1997.
[3] C. Garcia and M. Delakis. A neural architecture for fast and robust face detection. IEEE-IAPR
Int. Conference on Pattern Recognition, pages 40?43, 2002.
[4] F. J. Huang and Y. LeCun. Loss functions for discriminative training of energy-based graphical
models. Technical report, Courant Institute of Mathematical Science, NYU, June 2004.
[5] M. Jones and P. Viola. Fast multi-view face detection. Technical Report TR2003-96, Mitsubishi
Electric Research Laboratories, 2003.
[6] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document
recognition. Proceedings of the IEEE, 86(11):2278?2324, November 1998.
[7] S. Z. Li, L. Zhu, Z. Zhang, A. Blake, H. Zhang, and H. Shum. Statistical learning of multi-view
face detection. In Proceedings of the 7th European Conference on Computer Vision-Part IV,
2002.
[8] Y. Li, S. Gong, and H. Liddell. Support vector regression and classification based multi-view
face detection and recognition. In Face and Gesture, 2000.
[9] H. Moon and M. L. Miller. Estimating facial pose from sparse representation. In International
Conference on Image Processing, Singapore, 2004.
[10] A. Pentland, B. Moghaddam, and T. Starner. View-based and modular eigenspaces for face
recognition. In CVPR, 1994.
[11] H. A. Rowley, S. Baluja, and T. Kanade. Neural network-based face detection. PAMI, 20:22?38,
1998.
[12] H. A. Rowley, S. Baluja, and T. Kanade. Rotation invariant neural network-based face detection.
In Computer Vision and Pattern Recognition, 1998.
[13] H. Schneidermn and T. Kanade. A statistical method for 3d object detection applied to faces
and cars. In Computer Vision and Pattern Recognition, 2000.
[14] K. Sung and T. Poggio. Example-based learning of view-based human face detection. PAMI,
20:39?51, 1998.
[15] R. Vaillant, C. Monrocq, and Y. LeCun. Original approach for the localisation of objects in
images. IEE Proc on Vision, Image, and Signal Processing, 141(4):245?250, August 1994.
[16] P. Viola and M. Jones. Rapid object detection using a boosted cascade of simple features. In
Proceedings IEEE Conf. on Computer Vision and Pattern Recognition, pages 511?518, 2001.
| 2638 |@word multitask:1 version:1 advantageous:1 seems:1 mitsubishi:1 brightness:1 profit:1 initial:2 configuration:3 score:1 shum:1 ours:1 document:1 current:1 com:1 comparing:1 must:6 tilted:4 confirming:1 designed:5 update:6 half:3 fewer:1 selected:1 plane:10 parameterizations:2 location:3 arctan:1 zhang:2 mathematical:1 consists:1 combine:1 rapid:1 roughly:2 behavior:2 multi:10 detects:2 abound:1 estimating:1 underlying:1 interpreted:1 finding:4 vaillant:1 nj:2 sung:1 every:1 collecting:1 voting:1 classifier:1 wrong:1 before:1 positive:12 local:1 osadchy:2 consequence:1 pami:2 black:1 plus:3 collect:1 co:8 range:6 lecun:4 testing:2 definite:1 area:1 significantly:1 cascade:1 projection:1 pre:1 integrating:1 word:1 convenience:1 synergistic:1 applying:2 map:9 regardless:1 independently:1 resolution:1 x32:1 estimator:1 parameterizing:1 importantly:1 embedding:1 handle:2 variation:11 rita:1 recognition:8 lay:1 observed:1 bottom:1 parameterize:1 ensures:1 decrease:1 rowley:4 trained:14 solving:1 iapr:1 serve:1 upon:1 easily:2 various:3 america:2 train:5 fast:3 describe:4 detected:5 sc:5 tell:1 outside:1 whose:1 modular:1 larger:1 cvpr:1 say:1 s:5 otherwise:2 gi:2 advantage:1 net:3 interaction:1 product:1 relevant:1 flexibility:1 achieve:1 description:1 produce:1 incremental:1 rotated:2 object:3 derive:1 gong:1 pose:71 ij:2 rescale:1 implemented:2 c:6 involves:1 come:1 indicate:1 fij:1 closely:1 correct:2 attribute:1 annotated:2 radius:1 centered:1 human:3 generalization:1 proximity:1 blake:1 exp:4 mapping:1 claim:1 matthew:1 smallest:1 estimation:13 proc:1 integrates:2 label:6 tilting:1 minimization:3 hope:1 mit:2 boosted:1 surveillance:1 probabilistically:1 encode:2 derived:1 l0:4 june:1 properly:1 indicates:1 pentium:3 contrast:1 detect:4 sense:1 glass:1 inference:1 integrated:1 compatibility:1 pixel:4 classification:1 equal:3 having:1 cartoon:1 identical:2 jones:5 yaw:16 report:5 tr2003:1 employ:3 few:1 randomly:1 simultaneously:2 preserve:1 resulted:1 replaced:1 geometry:1 attempt:1 detection:45 highly:1 possibility:1 localisation:1 yielding:1 moghaddam:1 integral:1 closer:2 necessary:1 poggio:1 facial:5 eigenspaces:1 iv:1 accommodated:1 circle:3 desired:3 instance:1 column:1 increased:1 caruana:1 cost:1 subset:6 uniform:1 examining:1 iee:1 reported:2 answer:2 combined:2 lush:3 international:1 starner:1 probabilistic:2 together:5 again:1 choose:1 huang:1 conf:1 li:2 bold:1 int:1 satisfy:3 caused:1 depends:1 performed:2 view:17 lab:4 portion:1 start:1 sort:1 parallel:2 annotation:4 minimize:1 accuracy:2 moon:1 convolutional:10 roll:6 atan2:4 miller:2 yield:3 spaced:1 correspond:1 conceptually:1 raw:2 produced:1 regimen:1 mere:1 lighting:1 cc:5 published:2 classified:2 converged:2 simultaneous:2 detector:21 strongest:1 manual:3 against:7 energy:15 frequency:2 color:1 car:1 improves:1 focusing:1 courant:2 improved:1 done:1 though:1 box:3 just:1 hand:3 overlapping:1 building:2 effect:2 requiring:1 consisted:2 verify:1 normalized:1 inductive:1 analytically:1 symmetric:1 laboratory:1 undesired:1 gw:3 white:1 sin:5 during:1 cosine:2 complete:1 l1:4 image:41 novel:2 fi:1 respectively1:1 argminy:1 rotation:11 stepping:2 interpret:1 refer:1 gibbs:1 replicating:1 had:3 surface:1 operating:2 etc:1 curvature:2 closest:3 optimizing:2 discard:1 binary:2 kgw:3 came:1 swept:1 minimum:3 signal:1 semi:1 desirable:1 smooth:1 technical:2 match:2 gesture:1 sphere:1 posited:1 equally:1 pitch:5 basic:1 hair:1 regression:1 vision:5 cmu:2 publish:1 represent:1 tailored:1 want:2 separately:1 interval:1 finiteness:1 operate:1 pass:2 suspect:1 subject:1 ample:1 near:1 bengio:1 variety:2 switch:1 architecture:10 topology:1 idea:2 haffner:1 judiciously:1 shift:1 whether:3 motivated:1 expression:1 assist:1 york:1 nine:1 cause:1 hardware:1 simplest:1 argminz:1 http:1 percentage:4 singapore:1 shifted:1 estimated:7 per:8 correctly:4 group:2 key:1 threshold:5 nevertheless:1 drawn:1 localize:1 verified:1 graph:3 fraction:1 convert:2 run:2 angle:5 parameterized:2 schneiderman:2 almost:1 yann:2 p3:1 patch:2 decision:2 comparable:1 layer:4 quadratic:1 badly:1 x2:1 according:1 smaller:1 cun:1 making:1 s1:3 invariant:2 indexing:1 restricted:1 discus:1 count:1 know:1 end:2 apply:1 away:3 appropriate:1 disagreement:1 robustness:1 original:1 running:4 subsampling:2 ensure:1 include:2 graphical:1 exploit:2 build:2 webcam:1 lenet5:1 skin:1 intend:1 traditional:1 gradient:3 distance:4 separate:2 parametrized:1 manifold:22 collected:1 minimizing:2 difficult:1 mostly:1 negative:3 design:1 unknown:1 perform:1 november:1 pentland:1 viola:5 looking:1 frame:3 locate:2 august:1 inferred:1 trainable:2 security:1 hour:1 below:1 pattern:4 built:2 reliable:1 including:4 video:1 endto:1 mouth:1 suitable:1 natural:3 cascaded:1 zhu:1 representing:1 scheme:1 eye:1 identifies:1 axis:4 loss:9 localized:1 degree:4 pij:2 sufficient:3 s0:3 pi:1 last:1 bias:1 side:1 institute:2 wide:3 face:97 midpoint:2 sparse:1 tolerance:4 ghz:3 curve:3 dimension:2 benefit:2 world:1 crosshairs:1 author:1 made:2 regressors:1 replicated:2 counted:2 far:5 approximate:1 synergy:5 discriminative:3 grayscale:1 search:1 latent:1 table:3 kanade:5 learn:2 robust:5 mlm:1 bottou:2 european:1 electric:1 did:1 profile:10 roc:3 fashion:1 slow:1 decoded:1 sf:1 lie:1 third:2 down:2 formula:1 embed:1 specific:1 nyu:2 list:2 false:8 nec:4 clamping:1 depicted:1 garcia:1 likely:1 scalar:1 applies:1 corresponds:1 viewed:1 goal:1 consequently:1 considerable:2 experimentally:2 change:4 typical:2 baluja:2 total:1 gij:5 isomorphic:1 ew:46 indicating:1 select:1 internal:1 support:2 frontal:4 princeton:2 tested:1 |
1,804 | 2,639 | A Probabilistic Model for Online Document
Clustering with Application to Novelty Detection
Jian Zhang?
?School of Computer Science
Cargenie Mellon University
Pittsburgh, PA 15213
[email protected]
Zoubin Ghahramani??
? Gatsby Computational Neuroscience Unit
University College London
London WC1N 3AR, UK
[email protected]
Yiming Yang?
?School of Computer Science
Cargenie Mellon University
Pittsburgh, PA 15213
[email protected]
Abstract
In this paper we propose a probabilistic model for online document clustering. We use non-parametric Dirichlet process prior to model the growing number of clusters, and use a prior of general English language
model as the base distribution to handle the generation of novel clusters.
Furthermore, cluster uncertainty is modeled with a Bayesian Dirichletmultinomial distribution. We use empirical Bayes method to estimate
hyperparameters based on a historical dataset. Our probabilistic model
is applied to the novelty detection task in Topic Detection and Tracking
(TDT) and compared with existing approaches in the literature.
1
Introduction
The task of online document clustering is to group documents into clusters as long as
they arrive in a temporal sequence. Generally speaking, it is difficult for several reasons:
First, it is unsupervised learning and the learning has to be done in an online fashion,
which imposes constraints on both strategy and efficiency. Second, similar to other learning
problems in text, we have to deal with a high-dimensional space with tens of thousands of
features. And finally, the number of clusters can be as large as thousands in newswire data.
The objective of novelty detection is to identify the novel objects from a sequence of data,
where ?novel? is usually defined as dissimilar to previous seen instances. Here we are interested in novelty detection in the text domain, where we want to identify the earliest report
of every new event in a sequence of news stories. Applying online document clustering
to the novelty detection task is straightforward by assigning the first seed of every cluster
as novel and all its remaining ones as non-novel. The most obvious application of novelty
detection is that, by detecting novel events, systems can automatically alert people when
new events happen, for example.
In this paper we apply Dirichlet process prior to model the growing number of clusters, and
propose to use a General English language model as a basis of newly generated clusters.
In particular, the new clusters will be generated according to the prior and a background
General English model, and each document cluster is modeled using a Bayesian Dirichletmultinomial language model. The Bayesian inference can be easily carried out due to
conjugacy, and model hyperparameters are estimated using a historical dataset by the empirical Bayes method. We evaluate our online clustering algorithm (as well as its variants)
on the novelty detection task in TDT, which has been regarded as the hardest task in that
literature [2].
The rest of this paper is organized as follows. We first introduce our probabilistic model
in Section 2, and in Section 3 we give detailed information on how to estimate model
hyperparameters. We describe the experiments in Section 4, and related work in Section 5.
We conclude and discuss future work in Section 6.
2
A Probabilistic Model for Online Document Clustering
In this section we will describe the generative probabilistic model for online document
(x)
(x)
(x)
clustering. We use x = (n1 , n2 , . . . , nV ) to represent a document vector where each
(x)
element nv denotes the term frequency of the v th corresponding word in the document
x, and V is the total size of the vocabulary.
2.1
Dirichlet-Multinomial Model
The multinomial distribution has been one of the most frequently used language models for
modeling documents in information retrieval. It assumes that given the set of parameters
? = (?1 , ?2 , . . . , ?V ), a document x is generated with the following probability:
PV
V
(x)
nv )! Y n(x)
(
p(x|?) = QVv=1 (x)
?v v .
v=1 nv ! v=1
From the formula we can see the so-called naive assumption: words are assumed to be independent of each other. Given a collection of documents generated from the same model,
the parameter ? can be estimated with Maximum Likelihood Estimation (MLE).
In a Bayesian approach we would like to put a Dirichlet prior over the parameter (? ?
Dir(?)) such that the probability
R of generating a document is obtained by integrating over
the parameter space: p(x) = p(?|?)p(x|?)d?. This integration can be easily written
down due to the conjugacy between Dirichlet and multinomial distributions. The key difference between the Bayesian approach and the MLE is that the former uses a distribution
to model the uncertainty of the parameter ?, while the latter gives only a point estimation.
2.2
Online Document Clustering with Dirichlet Process Mixture Model
In our system documents are grouped into clusters in an online fashion. Each cluster is
modeled with a multinomial distribution whose parameter ? follows a Dirichlet prior. First,
a cluster is chosen based on a Dirichlet process prior (can be either a new or existing
cluster), and then a document is drawn from that cluster.
We use Dirichlet Process (DP) to model the prior distribution of ??s, and our hierarchical
model is as follows:
xi |ci ? M ul(.|? (ci ) )
?i
G
iid.
?
?
G
DP (?, G0 )
(1)
where ci is the cluster indicator variable, ? i is the multinomial parameter 1 for each document, and ? (ci ) is the unique ? for the cluster ci . G is a random distribution generated from
the Dirichlet process DP (?, G0 ) [4], which has a precision parameter ? and a base distribution G0 . Here our base distribution G0 is a Dirichlet distribution Dir(??1 , ??2 , . . . , ??V )
PV
with t=1 ?t = 1, which reflects our expected knowledge about G. Intuitively, our G0
distribution can be treated as the prior over general English word frequencies, which has
been used in information retrieval literature [6] to model general English documents.
The exact cluster-document generation process can be described as follows:
1. Let xi be the current document under processing (the ith document in the input
sequence), and C1 , C2 , . . . , Cm are already generated clusters.
2. Draw a cluster ci based on the following Dirichlet process prior [4]:
|C |
Pmj
p(ci = Cj ) =
(j = 1, 2, . . . , m)
? + j=1 |Cj |
p(ci = Cm+1 )
=
?+
?
Pm
j=1
(2)
|Cj |
Pm
where |Cj | stands for the cardinality of cluster j with j=1 |Cj | = i ? 1, and
with certain probability a new cluster Cm+1 will be generated.
3. Draw the document xi from the cluster ci .
2.3
Model Updating
Our models for each cluster need to be updated based on incoming documents. We can
write down the probability that the current document xi is generated by any cluster as
Z
p(xi |Cj ) = p(?(Cj ) |Cj )p(xi |?(Cj ) )d?(Cj ) (j = 1, 2, . . . , m, m + 1)
where p(? (Cj ) |Cj ) is the posterior distribution of parameters of the j th cluster (j =
1, 2, . . . , m) and we use p(? (Cm+1 ) |Cm+1 ) = p(? (Cm+1 ) ) to represent the prior distribution of the parameters of the new cluster for convenience. Although the dimensionality of
? is high (V ? 105 in our case), closed-form solution can be obtained under our Dirichletmultinomial assumption. Once the conditional probabilities p(xi |Cj ) are computed, the
probabilities p(Cj |xi ) can be easily calculated using the Bayes rule:
p(Cj )p(xi |Cj )
p(Cj |xi ) = Pm+1
0
0
j 0 =1 p(Cj )p(xi |Cj )
where the prior probability of each cluster is calculated using equation (2).
Now there are several choices we can consider on how to update the cluster models. The
first choice, which is correct but obviously intractable, is to fork m + 1 children of the
current system where the j th child is updated with document xi assigned to cluster j, while
the final system is a probabilistic combination of those children with the corresponding
probabilities p(Cj |xi ). The second choice is to make a hard decision by assigning the
current document xi to the cluster with the maximum probability:
p(Cj )p(xi |Cj )
ci = arg max p(Cj |xi ) = Pm+1
.
Cj
0
0
j 0 =1 p(Cj )p(xi |Cj )
For ? we use ?v to denote the v th element in the vector, ? i to denote the parameter vector that
generates the ith document, and ? (j) to denote the parameter vector for the j th cluster.
1
The third choice is to use a soft probabilistic updating, which is similar in spirit to the
Assumed Density Filtering (ADF) [7] in the literature. That is, each cluster is updated by
exponentiating the likelihood function with probabilities:
p(Cj |xi )
p(?(Cj ) |xi , Cj ) ?
p(xi |?(Cj ) )
p(?(Cj ) |Cj )
However, we have to specially deal with the new cluster since we cannot afford both timewise and space-wise to generate a new cluster for each incoming document. Instead, we
will update all existing clusters as above, and new cluster will be generated only if c i =
Cm+1 . We will use HD and PD (hard decision and probabilistic decision) to denote the
last two candidates in our experiments.
3
Learning Model Parameters
In the above probabilistic model there are still several hyperparameters not specified,
namely the ? and ? in the base distribution G0 = Dir(??1 , ??2 , . . . , ??V ), and the precision parameter ? in the DP (?, G0 ). Since we can obtain a partially labeled historical
dataset 2 , we now discuss how to estimate those parameters respectively.
We will mainly use the empirical Bayes method [5] to estimate those parameters instead
of taking a full Bayesian approach, since it is easier to compute and generally reliable
when the number of data points is relatively large compared to the number of parameters.
Because the ? i ?s are iid. from the random distribution G, by integrating out the G we get
X
?
1
?i |?1 , ?2 , . . . , ? i?1 ?
G0 +
? j
?+i?1
? + i ? 1 j<i ?
where the distribution is a mixture of continuous and discrete distributions, and the ? ?
denotes the probability measure giving point mass to ?.
Now suppose we have a historical dataset H which contains K labeled clusters H j (j =
1, 2, . . . , K), with the k th cluster Hk = {xk,1 , xk,2 , . . . , xk,mk } having mk documents.
The joint probability of ??s of all documents can be obtained as
p(?1 , ?2 , . . . , ? |H| ) =
|H|
Y
(
i=1
X
?
1
G0 +
? j)
?+i?1
? + i ? 1 j<i ?
where |H| is the total number of documents. By integrating over the unknown parameter
??s we can get
?
?
Z Y
|H|
? p(xi |?i )? p(?1 , ?2 , . . . , ? |H| )d?1 d?2 . . . d? |H|
p(H) =
i=1
=
?
Z
|H|
Y
? p(xi |?i )(
i=1
1
?
G0 +
?+i?1
?+i?1
X
j<i
?
??j )d?i ?
(3)
Empirical Bayes method can be applied to equation (3) to estimate the model parameters
by maximization3 . In the following we discuss how to estimate parameters individually in
detail.
2
Although documents are grouped into clusters in the historical dataset, we cannot make directly
use of those labels due to the fact that clusters in the test dataset are different from those in the
historical dataset.
3
Since only a subset of documents are labeled in the historical dataset H, the maximization is
only taken over the union of the labeled clusters.
Estimating ?t ?s
3.1
Our hyperparameter ? vector contains V number of parameters for the base distribution G 0 ,
which can be treated as the expected distribution of G ? the prior of the cluster parameter
??s.
Although ? contains V ? 105 number of actual parameters in our case, we can still use
the empirical Bayes to do a reliable point estimation since the amount of data we have to
represent general English is large (in our historical dataset there are around 10 6 documents,
around 1.8 ? 108 English words in total) and highly informative about ?. We use the
P
(H)
(H)
(H)
(H)
(x)
smoothed estimation ? ? (1 + n1 , 1 + n2 , . . . , 1 + nV ) where nt = x?H nt
PV
is the total number of times that term t happened in the collection H, and t=1 ?t should
be normalized to 1. Furthermore, the pseudo-count one is added to alleviate the out-ofvocabulary problem.
Estimating ?
3.2
Though ? is just a scalar parameter, it has the effect to control the uncertainty of the prior
knowledge about how clusters are related to the general English model with the parameter
?. We can see that ? controls how far each new cluster can deviate from the general English
model 4 . It can be estimated as follows:
K
K Z
Y
Y
?? = arg max
p(Hk |?) = arg max
p(Hk |?(k) )p(?(k) |?)d? (k)
(4)
?
?
k=1
k=1
?? can be numerically computed by solving the following equation:
K?(?) ? K
V
X
?(??v )?v +
v=1
K X
V
X
k)
?(??v + n(H
)?v ?
v
k=1 v=1
where the digamma function ?(x) is defined as ?(x) ?
K
X
?(? +
k=1
d
dx
V
X
k)
n(H
)=0
v
v=1
ln ?(x).
Alternatively we can choose ? by evaluating over the historical dataset. This is applicable
(though computationally expensive) since it is only a scalar parameter and we can precompute its possible range based on equation (4).
Estimating ?
3.3
The precision parameter ? of the DP is also very important for the model, which controls
how far the random distribution G can deviate from the baseline model G0 . In our case, it is
also the prior belief about how quickly new clusters will be generated in the sequence. Similarly we can use equation (3) to estimate ?, since items related to ? can be factored out as
Q|H| ?yi
L
i=1 ?+i?1 . Suppose we have a labeled subset H = {(x1 , y1 ), (x2 , y2 ), . . . , (xM , yM )}
of training data, where yi is 1 if xi is a novel document or 0 otherwise. Here we describe
two possible choices:
1. The simplest way is to assume that ? is a fixed constant during the process, and it
?y t
L
? = arg max? Q
can be computed as ?
i?H L ?+i?1 , here H denotes the subset of
indices of labeled documents in the whole sequence.
2. The assumption that ? is fixed maybe restrictive in reality, especially considering
the fact that it reflects the generation rate of new clusters. More generally, we
4
The mean and variance of a Dirichlet distribution (?1 , ?2 , . . . , ?V ) ? Dir(??1 , ??2 , . . . , ??V )
(1??v )
are: E[?v ] = ?v and Var[?v ] = ?v(?+1)
.
can assume that ? is some function of variable i. In particular, we assume ? =
a/i + b + ci where a, b and c are non-negative numbers. This formulation is a
generalization of the above case, where the i?1 term allows a much faster decrease
at the beginning, and c is the asymptotic rate of events happening as i ? ?.
Again the parameters a, b and c are estimated by MLE over the training dataset:
yi
Q
a
?, ?b, c? = arg maxa,b,c>0 i?H L (a/i+b+ci)
a/i+b+ci+i .
4
Experiments
We apply the above online clustering model to the novelty detection task in Topic Detection
and Tracking (TDT). TDT has been a research community since its 1997 pilot study, which
is a research initiative that aims at techniques to automatically process news documents in
terms of events. There are several tasks defined in TDT, and among them Novelty Detection
(a.k.a. First Story Detection or New Event Detection) has been regarded as the hardest task
in this area [2]. The objective of the novelty detection task is to detect the earliest report
for each event as soon as that report arrives in the temporal sequence of news stories.
4.1
Dataset
We use the TDT2 corpus as our historical dataset for estimating parameters, and use the
TDT3 corpus to evaluate our model 5 . Notice that we have a subset of documents in the
historical dataset (TDT2) for which events labels are given. The TDT2 corpus used for
novelty detection task consists of 62,962 documents, among them 8,401 documents are
labeled in 96 clusters. Stopwords are removed and words are stemmed, and after that there
are on average 180 words per document. The total number of features (unique words) is
around 100,000.
4.2
Evaluation Measure
In our experiments we use the standard TDT evaluation measure [1] to evaluate our results.
The performance is characterized in terms of the probability of two types of errors: miss
and false alarm (PM iss and PF A ). These two error probabilities are then combined into a
single detection cost, Cdet , by assigning costs to miss and false alarm errors:
Cdet = CM iss ? PM iss ? Ptarget + CF A ? PF A ? Pnon?target
where
1. CM iss and CF A are the costs of a miss and a false alarm, respectively,
2. PM iss and PF A are the conditional probabilities of a miss and a false alarm, respectively, and
3. Ptarget and Pnon?target is the priori target probabilities (Ptarget = 1 ?
Pnon?target ).
It is the following normalized cost that is actually used in evaluating various TDT systems:
(Cdet )norm =
Cdet
min(CM iss ? Ptarget , CF A ? Pnon?target )
where the denominator is the minimum of two trivial systems. Besides, two types of evaluations are used in TDT, namely macro-averaged (topic-weighted) and micro-averaged
5
Strictly speaking we only used the subsets of TDT2 and TDT3 that is designated for the novelty
detection task.
(story-weighted) evaluations. In macro-averaged evaluation, the cost is computed for every
event, and then the average is taken. In micro-averaged evaluation the cost is averaged over
all documents? decisions generated by the system, thus large event will have bigger impact
on the overall performance. Note that macro-averaged evaluation is used as the primary
evaluation measure in TDT. In addition to the binary decision ?novel? or ?non-novel?, each
system is required to generated a confidence score for each test document. The higher the
score is, the more likely the document is novel. Here we mainly use the minimum cost to
evaluate systems by varying the threshold, which is independent of the threshold setting.
4.3
Methods
One simple but effective method is the ?GAC-INCR? clustering method [9] with cosine
similarity metric and TFIDF term weighting, which has remained to be the top performing
system in TDT 2002 & 2003 official evaluations. For this method the novelty confidence
score we used is one minus the similarity score between the current cluster xi and its nearest
neighbor cluster: s(xi ) = 1.0 ? maxj<i sim(ci , cj ), where ci and cj are the clusters that
xi and xj are assigned to, respectively, and the similarity is taken to be the cosine similarity
between two cluster vectors, where the ltc TFIDF term weighting scheme is used to scale
each dimension of the vector. Our second method is to train a logistic regression model
which combines multiple features generated by the GAC-INCR method. Those features
not only include the similarity score used by the first method, but also include the size of
its nearest cluster, the time difference between the current cluster and the nearest cluster,
etc. We call this method ?Logistic Regression?, where we use the posterior probability
p(novelty|xi ) as the confidence score. Finally, for our online clustering algorithm we
choose the quantity s(xi ) = log p(C0 |xi ) as the output confidence score.
4.4
Experimental Results
Our results for three methods are listed in Table 1, where both macro-averaged and microaveraged minimum normalized costs are reported 6 . The GAC-INCR method performs
very well, so does the logistic regression method. For our DP results, we observed that
using the optimized ?? will get results (not listed in the table) that are around 10% worse
than using the ? obtained through validation, which might be due to the flatness of the
optimal function value as well as the sample bias of the clusters in the historical dataset 7 .
Another observation is that the probabilistic decision does not actually improve the hard
decision performance, especially for the ?var option. Generally speaking, our DP methods
are comparable to the other two methods, especially in terms of topic-weighted measure.
Table 1: Results for novelty detection on TDT3 corpus
Method
GAC-INCR
Logistic Regression
DP with ?f ix , HD
DP with ?var , HD
DP with ?f ix , PD
DP with ?var , PD
Topic-weighted Cost
COST (Miss, FA)
0.6945 (0.5614, 0.0272)
0.7027 (0.5732, 0.0264)
0.7054 (0.4737, 0.0473)
0.6901 (0.5789, 0.0227)
0.7054 (0.4737, 0.0473)
0.9025 (0.8772, 0.0052)
Story-weighted Cost
COST (Miss, FA)
0.7090 (0.5614, 0.0301)
0.6911 (0.5732, 0.0241)
0.7744 (0.5965, 0.0363)
0.7541 (0.5789, 0.0358)
0.7744 (0.5965, 0.0363)
0.9034 (0.8772, 0.0053)
6
In TDT official evaluation there is also the DET curve, which is similar in spirit to the ROC curve
that can reflects how the performance changes as the threshold varies. We will report those results in
a longer version of this paper.
7
It is known that the cluster labeling process of LDC is biased toward topics that will be covered
in multiple languages instead of one single language.
5
Related Work
Zaragoza et al. [11] applied a Bayesian Dirichlet-multinomial model to the ad hoc information retrieval task and showed that it is comparable to other smoothed language models.
Blei et al. [3] used Chinese Restaurant Processes to model topic hierachies for a collection of documents. West et al. [8] discussed the sampling techniques for base distribution
parameters in the Dirichlet process mixture model.
6
Conclusions and Future Work
In this paper we used a hierarchical probabilistic model for online document clustering.
We modeled the generation of new clusters with a Dirichlet process mixture model, where
the base distribution can be treated as the prior of general English model and the precision
parameter is closely related to the generation rate of new clusters. Model parameters are
estimated with empirical Bayes and validation over the historical dataset. Our model is
evaluated on the TDT novelty detection task, and results show that our method is promising.
In future work we would like to investigate other ways of estimating parameters and use
sampling methods to revisit previous cluster assignments. We would also like to apply our
model to the retrospective detection task in TDT where systems do not need to make decisions online. Though its simplicity, the unigram multinomial model has its well-known
limitation, which is the naive assumption about word independence. We also plan to explore richer but still tractable language models in this framework. Meanwhile, we would
like to combine this model with the topic-conditioned framework [10] as well as incorporate hierarchical mixture model so that novelty detection will be conditioned on some topic,
which will be modeled by either supervised or semi-supervised learning techniques.
References
[1] The 2002 topic detection & tracking task definition
http://www.nist.gov/speech/tests/tdt/tdt2002/evalplan.htm, 2002.
and
evaluation
plan.
[2] Allan, J., Lavrenko, V. & Jin, H. First story detection in tdt is hard. In Proc. of CIKM 2000.
[3] Blei, D., Griffiths, T., Jordan, M. & Tenenbaum, J. Hierarchical topic models and the nested
chinese restaurant process. Advances in Neural Information Processing Systems, 15, 2003.
[4] Ferguson, T. A Bayesian analysis of some nonparametric problems. Annals of Statistics, 1:209?
230, 1973.
[5] Gelman A., Carlin, J., Stern, H. & Rubin, D. Bayesian Data Analysis (2nd ed.). CHAPMAN &
HALL/CRC, 2003.
[6] Miller, D., Leek, T. & Schwartz, R. Bbn at trec 7: Using hidden markov models for information
retrieval. In TREC-7, 1999.
[7] Minka, T. A family of algorithms for approximate Bayesian inference. Ph.D. thesis, MIT, 2001.
[8] West, M., Mueller, P. & Escobar, M.D. Hierarchical priors and mixture models, with application
in regression and density estimation. In Aspects of Uncertainty: A tribute to D. V. Lindley, A.F.M.
Smith and P. Freeman, (eds.), Wiley, New York.
[9] Yang, Y., Pierce, T. & Carbonell, J. A Study on Retrospective and On-line Event Detection. In
Proc. of SIGIR 1998.
[10] Yang, Y., Zhang, J., Carnobell, J. & Jin, C. Topic-conditioned novelty detection. In Proc. of 8th
ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2002.
[11] Zaragoza, H., Hiemstra, D., Tipping, D. & Robertson, S. Bayesian extension to the language
model for ad hoc information retrieval. In Proc. SIGIR 2003.
| 2639 |@word version:1 norm:1 nd:1 c0:1 minus:1 contains:3 score:7 document:47 existing:3 current:6 nt:2 stemmed:1 assigning:3 dx:1 written:1 happen:1 informative:1 update:2 generative:1 item:1 xk:3 beginning:1 ith:2 smith:1 blei:2 detecting:1 lavrenko:1 zhang:3 stopwords:1 alert:1 c2:1 initiative:1 consists:1 combine:2 introduce:1 allan:1 expected:2 frequently:1 growing:2 freeman:1 automatically:2 gov:1 actual:1 pf:3 cardinality:1 considering:1 estimating:5 mass:1 cm:10 maxa:1 temporal:2 pseudo:1 every:3 uk:2 control:3 unit:1 schwartz:1 might:1 range:1 averaged:7 unique:2 union:1 area:1 empirical:6 word:8 integrating:3 confidence:4 griffith:1 zoubin:2 get:3 convenience:1 cannot:2 gelman:1 put:1 applying:1 www:1 straightforward:1 sigir:2 simplicity:1 factored:1 rule:1 regarded:2 hd:3 handle:1 updated:3 annals:1 target:5 suppose:2 exact:1 us:1 pa:2 element:2 robertson:1 expensive:1 updating:2 labeled:7 observed:1 fork:1 thousand:2 news:3 decrease:1 removed:1 pd:3 solving:1 efficiency:1 basis:1 easily:3 joint:1 htm:1 various:1 train:1 describe:3 london:2 effective:1 pnon:4 labeling:1 whose:1 richer:1 otherwise:1 statistic:1 final:1 online:14 obviously:1 hoc:2 sequence:7 ucl:1 propose:2 macro:4 cluster:60 generating:1 escobar:1 yiming:2 object:1 ac:1 nearest:3 school:2 sim:1 c:2 closely:1 correct:1 crc:1 generalization:1 alleviate:1 tfidf:2 strictly:1 extension:1 around:4 hall:1 seed:1 estimation:5 proc:4 applicable:1 label:2 individually:1 grouped:2 reflects:3 weighted:5 mit:1 aim:1 varying:1 earliest:2 likelihood:2 mainly:2 hk:3 digamma:1 sigkdd:1 baseline:1 detect:1 inference:2 mueller:1 ferguson:1 hidden:1 interested:1 arg:5 among:2 overall:1 priori:1 plan:2 integration:1 once:1 having:1 sampling:2 chapman:1 hardest:2 unsupervised:1 future:3 report:4 micro:2 maxj:1 n1:2 ltc:1 tdt:15 detection:25 highly:1 investigate:1 mining:1 evaluation:11 mixture:6 arrives:1 wc1n:1 mk:2 instance:1 modeling:1 soft:1 ar:1 assignment:1 maximization:1 cost:12 subset:5 reported:1 varies:1 dir:4 combined:1 density:2 international:1 probabilistic:12 ym:1 quickly:1 again:1 thesis:1 choose:2 worse:1 ad:2 closed:1 bayes:7 option:1 lindley:1 variance:1 miller:1 identify:2 bayesian:11 iid:2 ed:2 definition:1 frequency:2 minka:1 obvious:1 newly:1 dataset:16 pilot:1 knowledge:3 dimensionality:1 organized:1 cj:34 actually:2 adf:1 higher:1 tipping:1 supervised:2 formulation:1 done:1 though:3 evaluated:1 furthermore:2 just:1 logistic:4 effect:1 normalized:3 y2:1 former:1 assigned:2 deal:2 zaragoza:2 during:1 incr:4 cosine:2 performs:1 wise:1 novel:10 multinomial:7 discussed:1 numerically:1 mellon:2 pm:7 similarly:1 newswire:1 language:9 similarity:5 longer:1 etc:1 base:7 posterior:2 showed:1 certain:1 binary:1 yi:3 seen:1 minimum:3 novelty:18 semi:1 full:1 multiple:2 flatness:1 faster:1 characterized:1 long:1 retrieval:5 mle:3 bigger:1 impact:1 variant:1 regression:5 denominator:1 cmu:2 metric:1 represent:3 dirichletmultinomial:3 c1:1 background:1 want:1 addition:1 jian:2 biased:1 rest:1 specially:1 nv:5 spirit:2 jordan:1 call:1 yang:3 xj:1 restaurant:2 independence:1 carlin:1 det:1 ul:1 retrospective:2 speech:1 york:1 afford:1 speaking:3 generally:4 detailed:1 listed:2 covered:1 maybe:1 amount:1 nonparametric:1 ten:1 tenenbaum:1 ph:1 simplest:1 generate:1 http:1 notice:1 happened:1 revisit:1 neuroscience:1 estimated:5 per:1 cikm:1 write:1 discrete:1 hyperparameter:1 group:1 key:1 threshold:3 drawn:1 uncertainty:4 arrive:1 family:1 draw:2 decision:8 comparable:2 constraint:1 x2:1 generates:1 aspect:1 min:1 performing:1 relatively:1 designated:1 according:1 combination:1 precompute:1 ptarget:4 intuitively:1 taken:3 pmj:1 equation:5 conjugacy:2 ln:1 computationally:1 discus:3 count:1 leek:1 tractable:1 apply:3 hierarchical:5 denotes:3 assumes:1 clustering:12 dirichlet:16 remaining:1 cf:3 top:1 include:2 giving:1 restrictive:1 ghahramani:1 especially:3 chinese:2 objective:2 g0:11 already:1 added:1 quantity:1 parametric:1 strategy:1 primary:1 fa:2 dp:11 carbonell:1 topic:12 trivial:1 reason:1 toward:1 besides:1 gac:4 modeled:5 index:1 tdt2:4 difficult:1 tribute:1 negative:1 stern:1 unknown:1 observation:1 markov:1 nist:1 jin:2 y1:1 trec:2 smoothed:2 community:1 namely:2 required:1 specified:1 optimized:1 usually:1 xm:1 max:4 reliable:2 belief:1 ldc:1 event:11 treated:3 indicator:1 scheme:1 improve:1 carried:1 naive:2 text:2 prior:17 literature:4 deviate:2 discovery:1 asymptotic:1 generation:5 limitation:1 filtering:1 var:4 validation:2 imposes:1 rubin:1 story:6 last:1 soon:1 english:10 bias:1 neighbor:1 taking:1 curve:2 calculated:2 vocabulary:1 stand:1 evaluating:2 dimension:1 collection:3 exponentiating:1 historical:13 far:2 approximate:1 incoming:2 corpus:4 pittsburgh:2 conclude:1 assumed:2 xi:29 alternatively:1 continuous:1 reality:1 table:3 promising:1 meanwhile:1 domain:1 official:2 whole:1 hyperparameters:4 alarm:4 n2:2 child:3 x1:1 west:2 i:6 roc:1 fashion:2 gatsby:2 wiley:1 precision:4 pv:3 candidate:1 third:1 weighting:2 ix:2 formula:1 down:2 remained:1 unigram:1 cargenie:2 intractable:1 false:4 ci:15 bbn:1 pierce:1 conditioned:3 easier:1 likely:1 explore:1 microaveraged:1 happening:1 qvv:1 tracking:3 partially:1 scalar:2 nested:1 acm:1 hierachies:1 conditional:2 hard:4 change:1 miss:6 total:5 called:1 experimental:1 college:1 people:1 latter:1 dissimilar:1 incorporate:1 evaluate:4 |
1,805 | 264 | 298
Okamoto, Kawato, Ioui aod Miyake
Model Based Image Compression and
Adaptive Data Representation by
Interacting Filter Banks
SeiMiyake
Toshiaki Okamoto, Mitsuo Kawato, Toshio Ioui
ATR Auditory and
Visual Perception Research Laboratories
Sanpeidani, Inuidani. Seika-cho. Soraku-gun
Kyoto 619-02. Japan
NHK Science and
Technical Research Laboratories
1-10-11. Kinuta. Setagaya
Tokyo 157 ? Japan
Abstract
To achieve high-rate image data compression
while maintainig a high quality reconstructed
image, a good image model and an efficient
way to represent the specific data of each
image must be introduced. Based on the
physiological knowledge of multi - channel
characteristics and inhibitory interactions
between them in the human visual system,
a mathematically coherent parallel architecture
for image data compression which utilizes the
Markov random field Image model and
interactions between a vast number of filter
banks, is proposed.
1. Introduction
Data compression has been one of the most important
and active areas in information theory and computer science.
The goal of image coding is reducing the number of bits
in data representation as much as possible, and reconstructing
a faithful duplicate of the original image. In order to achieve
a high compression ratio while maintaining the high quality
Model Based Image Compression
of the reconstructed image, a good image model and an
efficient way to represent image data must be found. Based
on physiological knowledge of the human visual system, we
propose a mathematically coherent parallel architecture for
the image data compression, which utilizes a stochastic iInage
model and interactions between a vast number of filter banks.
2. Model based image compression and
dynamic spatial filtering
The process of reconstructing an original image from
compressed data is an ill-posed problem, since an infinite
number of original images lead to the same compressed data
and solutions to the inverse problem can not uniquely be
determined. The coupled Markov random field (MRF) image
model proposed by Geman and Geman is introduced to resolve
this ill-posedness. The mean field approximation of the MRF
is equivalent to a recurrent type neural network with the
Ljapunov function (see Koch. Marroquin and Yuille as a
special case where the form of the Ljapunov function is
predetermined). Correspondingly, a similar deterministic
framework of image compression in which the MRF is
replaced by the recurrent network, can be developed.
Further, even if a good MRF model is introduced for
a family of images, the data for each image must be known
in order to reconstruct it. In previous studies of image data
compression, representation of image data is fixed in each
schema. On the other hand, in this paper, an adaptive data
representation is proposed, tuned to each image by competion
and cooperation of a vast number of filter banks.
Fig. 1 shows a block diagram of the proposed
communication system. Procedures at the encoder side are
(1) partial partition and segmentation of the image by the
299
300
Okamoto, Kawato, Inui and Miyake
line process of the MRF which represents the image
discontinuity, (2) learning of energy parameters which uses
the line process to define the MRF model in each segmented
area of the image, (3) adaptive data representation of images
by cooperation and competition of a vast number of filter
banks. (4) Information about energy value parameters, the
types of selected filter and their outputs, and the line
processes is transmitted, through communication channel.
(5) Image reconstruction is carried out at the decorder site
by stochastic relaxation based on the aquired MRF model.
output from the selected filters, and the line process. These
procedures are explained in detail below.
1. The set of line processes represents discontinuities in
the 3-dimensional world such as occluding contours or
boundaries between different objects. It is not necessarily
closed, but it can posess a strong tendency to do so if the
MRF model is appropriately chosen. Based on this property,
the image can be partially segmented into several regions.
2. If we adopt the MRF image model, the occurrence
probability
n(w) of each configuration w is Gibbsian:
n(w)= exp{-U(w)/T}
Z
Furthermore, the energy U(w) can be expressed as a summation
of local potential Vc(w) , which depends on the configuration
only in the clique C.
U(w)= L Vc(w)
CeSc
Determination of the local energy Vc is equivalent to defining
a specific MRF model of the image. Determination of the
local energy is equivalent to assigning a real value VEo to
Model Based Image Compression
every possible configuration within the clique C. These energy
parameters are estimated so that the Kullback divergence
G between the real image distribution P and the model image
distribution P' is minimized:
P(w)
G(V)=~P(w)log{p'(w. Vd}
The following learning equation can be derived In
approximately the same way as the learning rule of the
Boltzmann machine (Ackley, Hinton, Sejnowski. 1985).
Here L(C)
IS the characteristic function of the specific
configuration 'i of the clique C, that is. MC)=l if{Ys;sEC}='i
otherwise, I;(C)=O.
The first term on the right side is the
average number of configurations in the real image. The
second term on the right side is the average number of each
configuration
generated in the MRF with the energy Vc
when part of the image configuration is fixed to the given
'I
image.
3. This
procedure
is
based
on
the
multi - channel
characteristics of the human visual system. inhibitory
interaction between X-cell and Y -cell systems. and interactions
between columns with different orientation selectivity. etc.
We prepare a vast number of filters centered at each site
s in a variety of sizes. shapes and orientations. In particular)
we use two-dimensional Gaussian filters Gs(w) to represent
the DC components (i.e. average luminance) of the gray
level, and use the first-order derivative of the Gaussian filters
VGs(w) to represent the gradient of the gray levels. The filters
whose receptive fields significantly intersect with the line
process are inhibited. Inhibitory interactions between filters
of similar, shape and orientation at nearby sites are introduced
301
302
Okamoto, Kawato, Inui and Miyake
as well as self excitation to find the N-maximum outputs
of 'YGs, and to find the N-minimum outputs of the Laplacian
Gaussian ~Gs. Of course, 2N must be less than the number
of sites to attain data compression.
4. We transmit the local potential energy, the site of the
line process, and the outputs from the N - maximum, and
the outputs from the N Gaussain filters which correspond
to the N - minimum Laplacian Gaussain filters.
5. Image reconstruction is carried out by the usual
stochastic relaxation, that is, energy minimization with
simulated annealing. However, because we have data
constraints as output from the 2N selected filters, we need
to minimize the sum of the MRF model energy and the data
constraint energy:
If we do not further
compress the filter outputs, the
regularization parameter is increased to infinity during
constrained stochastic relaxation.
3. Experimental results
First, we ascertained that the proposed energy learning
rule works well for various images. Here, we report only
one example from the data compression experiments. We used
the shown in Fig. 2a to examine the potential of our scheme.
The image data consists of 256 pixels, each of which has
8 bit gray levels. We used the dynamic spatial sampling
of filter banks. Fig 2a also shows selected sample points
in the image as black dots, as well as a few examples of
selected filter shapes. Note that not only the density of
the sampling points, but also the selected filter shapes are
\
Model Based Image Compression
appropriate local characteristics of the image. Fig. 2b shows
the reconstructed image after 20 iterations of the relaxation
computation. The signal to noise ratio of the reconstructed
images was a bou t 38dB.
References
D. H., Ackley, G. E. Hinton, and T. J. Sejnowski,: "A Learning
algorithm for Boltzmann Machines.", Cognitive Science, vol.
9, pp.147 - 169, (1985).
S. Geman and D. Geman,: "Stochastic relaxation, Gibbs
distribution, and the Basian restoration of images", IEEE Trans.
vol. PAMI - 6, pp.721 - 741, (1984).
S. Hongo, M. Kawato, T. Inui, and S. Miyake,; "Contour
extraction of images on parallel computer", Proc. of 1th IJCNN,
(1989).
T. Inui, M. Kawato and R. Suzuki: "The mechanism of mental
scannIng In foveal vision", BioI. Cybern. vol. 30, pp.147 - 155,
(1978).
C. Koch, J. Marroquin, and A. Yuille: "Analog 'neural'
networks in early vision", Proc. Natl. Acad. Sci. USA, vol. 83,
pp.4263 - 4267, (1986).
303
I
Encoder
I
Decoder]
Image Model
Line Process
~
original
Image
.
"'-i
,~
Line
Process
Sampled
Dynamic Spatial
Filtering
1-+ Image Data
I
Sparse
IL __
S ~R l '
_l.E'i _
I
.JI
MRF
Parameters
1
r------?
~
~
~
~
~
'C
....=
~
=
=
o
~
-
r - - - - - - -
I Stochastic I
I Parameter :
I Estimation I
L _______ I
.
:
~
~elaxationl
______
- - - -- - -- - --
Image Model
I
I
I
I
I
I
I
Line
Process
~
Sampled
Image Data
MRF
Parameters
???
~
Reconstructed
Image
????
??
r--------,
:constrained :
I Relaxation I
L _______ ..J
- - - - - - J
MRF
~------. _Parameters
~
~
~
~
o
'0
S
~
~
o
~
Q
~
Fig. 1 Model Based Communication System
Model Based Image Compression
(a)
sampled data points
and filters
(b)
reconstructed Image
Fig. 2 Computer simulation
of image data compression
305
PART IV:
OPTIMIZATION AND CONTROL
| 264 |@word compression:16 sanpeidani:1 regularization:1 tokyo:1 laboratory:2 filter:20 simulation:1 stochastic:6 vc:4 centered:1 human:3 receptive:1 during:1 self:1 uniquely:1 usual:1 gradient:1 excitation:1 atr:1 simulated:1 vd:1 sci:1 configuration:7 foveal:1 decoder:1 gun:1 tuned:1 summation:1 mathematically:2 koch:2 image:57 ratio:2 assigning:1 exp:1 must:4 kawato:6 partition:1 predetermined:1 shape:4 adopt:1 early:1 ji:1 estimation:1 proc:2 analog:1 boltzmann:2 selected:6 prepare:1 gibbs:1 markov:2 gaussain:2 minimization:1 mental:1 defining:1 nhk:1 hinton:2 gaussian:3 communication:3 dot:1 dc:1 interacting:1 etc:1 posedness:1 introduced:4 consists:1 derived:1 inuidani:1 inui:4 selectivity:1 coherent:2 seika:1 examine:1 multi:2 discontinuity:2 trans:1 transmitted:1 minimum:2 below:1 perception:1 resolve:1 signal:1 pixel:1 ill:2 orientation:3 kyoto:1 segmented:2 technical:1 determination:2 spatial:3 special:1 developed:1 constrained:2 field:4 scheme:1 extraction:1 y:1 sampling:2 laplacian:2 represents:2 every:1 mrf:15 carried:2 vision:2 coupled:1 minimized:1 report:1 iteration:1 represent:4 control:1 duplicate:1 inhibited:1 few:1 veo:1 cell:2 divergence:1 local:5 annealing:1 diagram:1 replaced:1 cesc:1 acad:1 appropriately:1 filtering:2 ygs:1 mitsuo:1 approximately:1 pami:1 black:1 db:1 usa:1 bank:6 natl:1 course:1 variety:1 faithful:1 gibbsian:1 cooperation:2 architecture:2 block:1 partial:1 side:3 ascertained:1 procedure:3 iv:1 correspondingly:1 intersect:1 area:2 sparse:1 significantly:1 attain:1 boundary:1 world:1 increased:1 column:1 contour:2 suzuki:1 adaptive:3 soraku:1 restoration:1 cybern:1 reconstructed:6 equivalent:3 deterministic:1 kullback:1 hongo:1 clique:3 active:1 bou:1 miyake:4 scanning:1 cho:1 rule:2 inhibitory:3 density:1 estimated:1 channel:3 vgs:1 vol:4 transmit:1 necessarily:1 us:1 cognitive:1 derivative:1 luminance:1 vast:5 relaxation:6 noise:1 geman:4 japan:2 potential:3 sum:1 ackley:2 inverse:1 fig:6 coding:1 sec:1 site:5 region:1 family:1 depends:1 utilizes:2 closed:1 schema:1 bit:2 parallel:3 dynamic:3 g:2 minimize:1 il:1 ijcnn:1 specific:3 constraint:2 yuille:2 characteristic:4 infinity:1 correspond:1 physiological:2 nearby:1 various:1 aod:1 mc:1 sejnowski:2 whose:1 reconstructing:2 posed:1 visual:4 energy:12 pp:4 reconstruct:1 compressed:2 encoder:2 otherwise:1 expressed:1 partially:1 explained:1 okamoto:4 sampled:3 auditory:1 equation:1 knowledge:2 bioi:1 goal:1 propose:1 reconstruction:2 interaction:6 segmentation:1 marroquin:2 mechanism:1 infinite:1 determined:1 reducing:1 achieve:2 appropriate:1 ioui:2 competition:1 furthermore:1 occurrence:1 tendency:1 experimental:1 hand:1 occluding:1 original:4 compress:1 object:1 recurrent:2 aquired:1 pose:1 quality:2 gray:3 maintaining:1 strong:1 |
1,806 | 2,640 | A Three Tiered Approach for Articulated Object
Action Modeling and Recognition
Le Lu
Gregory D. Hager
Department of Computer Science
Johns Hopkins University
Baltimore, MD 21218
lelu/[email protected]
Laurent Younes
Center of Imaging Science
Johns Hopkins University
Baltimore, MD 21218
[email protected]
Abstract
Visual action recognition is an important problem in computer vision.
In this paper, we propose a new method to probabilistically model and
recognize actions of articulated objects, such as hand or body gestures,
in image sequences. Our method consists of three levels of representation. At the low level, we first extract a feature vector invariant to scale
and in-plane rotation by using the Fourier transform of a circular spatial
histogram. Then, spectral partitioning [20] is utilized to obtain an initial
clustering; this clustering is then refined using a temporal smoothness
constraint. Gaussian mixture model (GMM) based clustering and density
estimation in the subspace of linear discriminant analysis (LDA) are then
applied to thousands of image feature vectors to obtain an intermediate
level representation. Finally, at the high level we build a temporal multiresolution histogram model for each action by aggregating the clustering
weights of sampled images belonging to that action. We discuss how this
high level representation can be extended to achieve temporal scaling invariance and to include Bi-gram or Multi-gram transition information.
Both image clustering and action recognition/segmentation results are
given to show the validity of our three tiered representation.
1 Introduction
Articulated object action modeling, tracking and recognition has been an important research issue in computer vision community for decades. Past approaches [3, 13, 4, 6, 23, 2]
have used many different kinds of direct image observations, including color, edges, contour or moments [14], to fit a hand or body?s shape model and motion parameters.
In this paper, we propose to learn a small set of object appearance descriptors, and then
to build an aggregated temporal representation of clustered object descriptors over time.
There are several obvious reasons to base gesture or motion recognition on a time sequence
of observations. First, most hand or body postures are ambiguous. For example, in American Sign Language, ?D? and ?G?, ?H? and ?U? have indistinguishable appearance from some
viewpoints. Furthermore, these gestures are difficult to track from frame to frame due to
motion blur, lack of features, and complex self-occlusions. By modeling hand/body gesture
as a sequential learning problem, appropriate discriminative information can be retrieved
and more action categories can be handled.
In related work, Darrell and Pentland [7] describe dynamic time warping (DTW) to align
and recognize a space-time gesture against a stored library. To build the library, key views
are selected from incoming an video by choosing views that have low correlation with all
current views. This approach is empirical and does not guarantee any sort of global consistency of the chosen views. As a result, recognition may be unstable. In comparision, our
method describes image appearances uniformly and clusters them globally from a training
set containing different gestures.
For static hand posture recognition, Tomasi et al. [24] apply vector quantization methods
to cluster images of different postures and different viewpoints. This is a feature-based
approach, with thousands of features extracted for each image. However, clustering in a
high dimensional space is very difficult and can be unstable. We argue that fewer, more
global features are adequate for the purposes of gesture recognition. Furthermore, the
circular histogram representation has adjustable spatial resolution to accomodate differing
appearance complexities, and it is translation, rotation, and scale invariant.
In other work, [27, 9] recognize human actions at a distance by computing motion information between images and relying on temporal correlation on motion vectors across
sequences. Our work also makes use of motion information, but does not rely exclusively
on it. Rather, we combine appearance and motion cues to increase sensitivity beyond what
either can provide alone. Since our method is based on the temporal aggregation of image
clusters as a histogram to recognize an action, it can also be considered to be a temporal texton-like method [17, 16]. One advantage of the aggregated histogram model in a
time-series is that it is straightforward to accommodate temporal scaling by using a sliding
window. In addition, higher order models corresponding to bigrams or trigrams of simpler
?gestemes? can also be naturally employed to extend the descriptive power of the method.
In summary, there are four principal contributions in this paper. First, we propose a new
scale/rotation-invariant hand image descriptor which is stable, compact and representative.
Second, we introduce a methods for sequential smoothing of clustering results. Third,
we show LDA/GMM with spectral partitioning initialization is an effective way to learn
well-formed probability densities for clusters. Finally, we recognize image sequences as
actions efficiently based on a flexible histogram model. We also discuss improvement to
the method by incorporating motion information.
2 A Three Tiered Approach
We propose a three tiered approach for dynamic action modeling comprising low level
feature extraction, intermediate level feature vector clustering and high level histogram
recognition as shown in Figure 1.
Low Level: Rotation Invariant Feature
Extraction
Probabilistic Foreground Map
(GMM Color Segmentation,
Probabilistic Appearance Modeling,
Dynamic Texture Segmentation by
GPCA)
Intermediate Level: Clustering Presentation
for Image Frames
High Level: Aggregated Histogram Model
for Action Recognition
GMM/LDA Density Modeling for
Clusters
Temporal Aggregated
Multiresolution Histogram for
Action
Feature Extraction
via Circular/Fourier Representation
Temporally Constrained
Clustering (Refinement)
From Unigram to Bigram,
Multigram Histogram Model
Feature Dimension Reduction
via Variance Analysis
Framewise Clustering via
Spectral Segmentation
(Initialization)
Temporal Pyramids and Scaling
Figure 1: Diagram of a three tier approach for dynamic articulated object action modeling.
(a)
(b)
(c)
Figure 2: (a) Image after background subtraction (b) GMM based color segmentation
(c) Circular histogram for feature extraction.
2.1
Low Level: Rotation Invariant Feature Extraction
In the low level image processing, our goals are to locate the region of interest in an image and to extract a scale and in-plane rotation invariant feature vector as its descriptor.
In order to accomplish this, a reliable and stable foreground model of the target in question is expected. Depending on the circumstances, a Gaussian mixture model (GMM) for
segmentation [15], probabilistic appearance modeling [5], or dynamic object segmentation
by Generalized Principal Component Analysis (GPCA) [25] are possible solutions. In this
paper, we apply a GMM for hand skin color segmentation.
We fit a GMM by first performing a simple background subtraction to obtain a noisy foreground containing a hand object (shown in Figure 2 (a)). From this, more than 1 million RGB pixels are used to train skin and non-skin color density models with 10 Gaussian kernels for each class. Having done this, for new images a probability density ratio
Pskin /Pnonskin of these two classes is computed. If Pskin /Pnonskin is larger than 1, the
pixel is considered as skin (foreground) and is otherwise background. A morphological
operator is then used to clean up this initial segmentaion and create a binary mask for the
hand object. We then compute the centroid and second central moments of this 2D mask.
A circle is defined about the target by setting its center as the centroid and its radius as
2.8 times largest eigenvalues of the second central moment matrix (covering over 99% skin
pixels in Figure 2 (c)). This circle is then divided to have 6 concentric annuli which contain 1, 2, 4, 8, 16, 32 bins from inner to outer, respectively. Since the position and size of
this circular histogram is determined by the color segmentation, it is translation and scale
invariant.
We then normalize the density value Pskin + Pnonskin = 1 for every pixel within the
foreground mask (Figure 2) over the hand region. For each bin of the circular histogram,
we calculate the mean of Pskin ( ?log(Pskin ), or ?log(Pskin /Pnonskin ) are also possible
choices) of pixels in that bin as its value. The values of all bins along each circle form a
vector, and 1D Fourier transform is applied to this vector. The power spectra of all annuli
are ordered into a linear list producing a feature vector f~(t) of 63 dimensions representing
the appearance of a hand image.1 Note that the use of the Fourier power spectrum of the
annuli makes the representation rotation invariant.
2.2
Intermediate Level: Clustering Presentation for Image Frames
After the low level processing, we obtain a scale and rotation invariant feature vector as
an appearance representation for each image frame. The temporal evolution of feature
vectors represent actions. However, not all the images are actually unique in appearance.
1
An optional dimension reduction of feature vectors can be achieved by eliminating dimensions
which have low variance. It means that feature values of those dimensions do not change much in the
data, therefore are non-informative.
At the intermediate level, we cluster images from a set of feature vectors. This frame-wise
clustering is critical for dimension reduction and the stability of high level recognition.
Initializing Clusters by Spectral Segmentation There are two critical problems with
clustering algorithms: determining the true number of clusters and initializing each cluster. Here we use a spectral clustering method [20, 22, 26, 18] to solve both problems.
We first build the affinity matrix of pairwise distances between feature fectors2 . We then
perform a singular value decomposition on the affinity matrix with proper normalization
[20]. The number of clusters is determined by choosing the n dominant eigenvalues. The
corresponding eigenvectors are taken as an orthogonal subspace for all the data.
To get n cluster centers, we take the approach of [20] and choose vectors that minimize the
absolute value of cosine between any two cluster centers:
rand(0, N ) :
k=1
Pk?1
ID(k) =
(1)
n
n
~
~
arg mint=1..N c=1 | cos(f (ID(c)), f (t))| : n ? k > 1
where f~n (t) is the feature vector of image frame t after numerical normalization in [20]
and ID(k) is the image frame number chosen for the center of cluster k. N is the number of
images used for spectral clustering. For better clustering results, multiple restarts are used
for initialization.
Unlike [18], we find this simple clustering procedure is sufficient to obtain a good set of
clusters from only a few restarts. After initialization, the Kmeans [8] is used to smooth the
centers. Let C1 (t) denote the class label for image t, and ~g (c) = f~(ID(c)); c = 1 . . . n
denote cluster centers.
Refinement: Temporally Constrained Clustering Spectral clustering methods are designed for an unordered ?bag? of feature vectors, but, in our case, the temporal ordering
of image is an important source of information. In particular, the stablity of appearance is
easily computed by computing the motion energy3 between two frames. Let M (t) denote
the motion energy
between frames t and t?1. Define Tk,j = {t|C1 (t) = k, C1 (t?1) = j}
? (k, j) = P
and M
t?Tk,j M (t)/|Tk,j |. We now create a regularized clustering cost function
as
?
?
kg(c)?g(C2 (t?1))k
?
? e?kf (t)?g(c)k
M (t)
e?
+
?
(2)
C2 (t) = arg maxc=1..n Pn
kg(c)?g(C2 (t?1))k
Pn
?
? M
? c=1 e?kf (t)?g(c)k
? (c,C (t?1))
2
e
c=1
where ? is the weighting parameter. Here motion energy M (t) plays a role as the temperature T in simulated annealing. When it is high (strong motion between frames), the
motion continuity condition is violated and the labels of successive frames can change
freely; when it is low, the smoothness term constrains the possible transitions of classes
? (k, j).
with low M
With this in place, we now scan through the sequence searching for C2 (t) of maximum
value given C2 (t ? 1) is already fixed. 4 This temporal smoothing is most relevant with
images with motions, and static frames are already stably clustered and therefore their
cluster labels to not change.
2
The exponent of either Euclidean distance or Cosine distance between two feature vectors can
be used in this case.
3
A simple method is to compute motion energy as the Sum of Squared Differences (SSD) by
subtracting two Pskin density masses from successive images.
4
? (k, j) changes after scanning the labels of the image sequence once, thus more
Note that M
iterations could be used to achieve more accurate temporal smoothness of C3 (t), t = 1..N . From
our experiments, more iterations does not change the result much.
GMM for Density Modeling and Smoothing Given clusters, we build a probability density model for each. A Gaussian Mixture Model [11, 8] is used to gain good local relaxation
based on the initial clustering result provided by the above method and good generalization
for new data. Due to the curse of dimensionality, it is difficult to obtain a good estimate
of a high dimensional density function with limited and largely varied training data. We
introduce an iterative method incorporating Linear Discriminative Analysis (LDA) [8] and
a GMM in an EM-like fashion to perform dimensional reduction. The initial clustering
labels help to build the scatter matrices for LDA. The optimal projection matrix of LDA is
then obtained from the decomposition of clusters? scatter matrices [8]. The original feature
vectors can be further projected into a low dimensional space, which improves the estimation of multi-variate Gaussian density function. With the new clustering result from GMM,
LDA?s scatter matrices and projection matrix can be re-estimated, and GMM can also be
re-modeled in the new LDA subspace. This loop converges within 3 ? 5 iterations from
our experiments. Intuitively, LDA projects the data into a low dimensional subspace where
the image clusters are well separated, which helps to have a good parameter estimation for
GMM with limited data. Given more accurate GMM, more accurate clustering results are
obtained, which also causes better estimate of LDA. The theoretical proof of convergence
is undertaken. After this process, we have a Gaussian density model for each cluster.
2.3
High Level: Aggregated Histogram Model for Action Recognition
Given a set of n clusters, define w(t) = [pc1 (f (t)), pc2 (f (t)), ..., pcn (f (t))]T where px (y)
denotes the density value of the vector y with respect to the GMM for cluster x. An action
is then a trajectory of [w(t1 ), w(t1 + 1), ..., w(t2 )]T in <n . For recognition purposes, we
want to calculate some discriminative statistics from each trajectory. One natural way is to
Pt2
w(t)/(t2 ? t1 + 1) over time which is a temporal weighted
use its mean Ht1 ,t2 = t=t
1
histogram. Note that the histogram Ht1 ,t2 bins are precisely corresponding to the trained
clusters.
From the training set, we aggregate the cluster weights of images within a given hand action
to form a histogram model. In this way, a temporal image sequence corresponding to one
action is represented by a single vector. The matching of different actions is equivalent to
compute the similarity of two histograms which has variants. Here we use Bhattacharyya
similarity metric [1] which has has several useful properties including: it is an approximation of ?2 test statistics with fixed bias; it is self-consistent; it does not have the singularity
problem while matching empty histogram bins; and its value is properly bounded within
?
[0, 1]. Assume we have a library of action histograms H1? , H2? , ..., HM
, the class label of a
? t1 ,t2 is determined by the following equation.
new action H
?
# 12 ?
"
n q
?
?
X
? t1 ,t2 ) = arg min
? t1 ,t2 (c)
? t1 ,t2 ) = 1 ?
(3)
L(H
D(Hl? , H
Hl? (c) ? H
l=1..M ?
?
c=1
This method is low cost because only one exemplar per action category is needed.
One problem with this method is that all sequence information has been compressed, e.g.,
we cannot distinguish an opening hand gesture from a closing hand using only one histogram. This problem can be easily solved by subdividing the sequence and histogram
model into m parts: Htm1 ,t2 = [Ht1 ,(t1 +t2 )/m , ..., H(t1 +t2 )?(m?1)/m,t2 ]T . For an extreme
case when one frame is a subsequence, the histogram model simply becomes exactly the
vector form of the representative surface.
We intend to classify hand actions with speed differences into the same category. To
achieve this, the image frames within a hand action can be sub-sampled to build a set
of temporal pyramids. In order to segment hand gestures from a long video sequence, we
create several sliding windows with different frame sampling rates. The proper time scaling
magnitude is found by searching for the best fit over temporal pyramids.
Taken together, the histogram representation achieves an adjustable multi-resolution measurement to describe actions. A Hidden Markov Model (HMM) with discrete observations
could be also employed to train models for different hand actions, but more template samples per gesture class are required. The histogram recognition method has the additional
advantage that it does not depend on extremely accurate frame-wise clustering. A small
proportion of incorrect labels does not effect the matching value much. In comparison,
in an HMM with few training samples, outliers seriously impact the accuracy of learning.
From the viewpoint of considering hand actions as a language process, our model is an integration of individual observations (by labelling each frame with a set of learned clusters)
from different time slots. The labels? transitions between successive frames are not used
to describe the temporal sequence. By subdividing the histogram, we are extending the
representation to contain bigram, trigram, etc. information.
3 Results
We have tested our three tiered method on the problem of recognizing sequences of hand
spelling gestures.
Framewise clustering. We first evaluate the low level representation of single images and
intermediate clustering algorithms. A training set of 3015 images are used. The frame-toframe motion energy is used to label images as static or dynamic. For spectral clustering,
3 ? 4 restarts from both the dynamic and static set are sufficient to cover all the modes in
the training set. Then, temporal smoothing is employed and a Gaussian density is calculated
for each cluster in a 10 dimensional subspace of the LDA projection. As a result, 24 clusters
are obtained which contain 16 static and 8 dynamic modes. Figure 3 shows 5 frames
closest to the mean of the probability density of cluster 1, 3, 19, 5, 13, 8, 21, 15, 6, 12. It
can be seen that clustering results are insensitive to artifacts of skin segmentation. From
Figure 3, it is also clear that dynamic modes have significantly larger determinants than
static ones. The study of the eigenvalues of covariance matrices shows that their superellipsoid shapes are expanded within 2 ? 3 dimensions or 6 ? 8 dimensions for static
or dynamic clusters. Taken together, this means that static clusters are quite tight, while
dynamic clusters contain much more in-class variation. From Figure 4 (c), dynamic clusters
gain more weight during the smoothing process incorporating the temporal constraint and
subsequent GMM refinement.
Figure 3: Image clustering results after low and intermediate level processing.
Action recognition and segmentation. For testing images, we first project their feature
A
A
YA
YA
W
A
XA
Y
A
Y
A
W
A
X
A
(a)
(b)
(c)
(d)
Figure 4: (a) Affinity matrix of 3015 images. (b) Affinity matrices of cluster centoids (from upper
left to lower right) after spectral clustering, temporal smoothing and GMM. (c) Labelling results of
3015 images (red squares are frames whose labels changed with smoothing process after spectral
clustering). (d) The similarity matrix of segmented hand gestures. The letters are labels of gestures.
vectors into the LDA subspace. Then, the GMM is used to compute their weights with
respect to each cluster. We manually choose 100 sequences for testing purposes, and compute their similarities with respect to a library of 25 gestures. The length of the action
sequences was 9 ? 38 frames. The temporal scale of actions in the same category ranged
from 1 to 2.4. The results were recognition rates of 90% and 93% without/with temporal
smoothing (Equation 2). Including the top three candidates, the recognition rates increase
to 94% and 96%, respectively. We also used the learned model and a sliding window with
temporal scaling to segment actions from a 6034 frame video sequence containing dynamic
gestures and static hand postures. The similarity matrices among 123 actions found in the
video is shown in Figure 4 (d). 106 out of 123 actions (86.2%) are correctly segmented and
recognized.
Integrating motion information. As noted previously, our method cannot distinguish
opening/closing hand gestures without temporally subdividing histograms. An alternative
solution is to integrate motion information5 between frames. Motion feature vectors are
also clustered, which results a joint (appearance and motion) histogram model for actions.
We assume independence of the data and therefore simple contatenate these two histograms
into a single action representation. From our preliminary experiments, both motion integration and histogram subdivision are comparably effective to recognize gestures with
opposite direction.
4 Conclusion and Discussion
We have presented a method for classifying the motion of articulated gestures using
LDA/GMM-based clustering methods and a histogram-based model of temporal evolution.
Using this model, we have obtained extremely good recognition results using a relatively
coarse representation of appearance and motion in images.
There are mainly three methods to improve the performance of histogram-based classification, i.e., adaptive binning, adaptive subregion, and adaptive weighting [21]. In our
approach, adaptive binning of the histogram is automatically learned by our clustering
algorithms; adaptive subregion is realized by subdividing action sequences to enrich the
histogram?s descriptive capacity in the temporal domain; adaptive weighting is achieved
from the trained weights of Gaussian kernels in GMM.
Our future work will focus on building a larger hand action database containing 50 ? 100
5
Motion information can be extracted by first aligning two hand blobs, subtracting two skin-color
density masses, then using the same circular histogram in section 2.1 to extract a feature vector for
positive and negative density residues respectively. Another simple way is to subtract two frames?
feature vectors directly.
categories for more extensive testing, and on extending the representation to include other
types of image information (e.g. contour information). Also, by finding an effective foreground segmentation module, we intend to apply the same methods to other applications
such as recognizing stylized human body motion.
References
[1] F. Aherne, N. Thacker, and P. Rockett, The Bhattacharyya Metric as an Absolute Similarity
Measure for Frequency Coded Data, Kybernetika, 34:4, pp. 363-68, 1998.
[2] V. Athitsos and S. Sclaroff, Estimating 3D Hand Pose From a Cluttered Image, CVPR, 2003.
[3] M. Brand, Shadow Puppetry, ICCV, 1999.
[4] R. Bowden and M. Sarhadi, A Non-linear of Shape and Motion for Tracking Finger Spelt
American Sign Language, Image and Vision Computing, 20:597-607, 2002.
[5] T. Cootes, G. Edwards and C. Taylor, Active Appearance Models, IEEE Trans. PAMI, 23:6, pp.
681-685, 2001.
[6] D. Cremers, T. Kohlberger and C. Schnrr, Shape statistics in Kernel Space for Variational Image
Segmentation, Pattern Recognition, 36:1929-1943, 2003.
[7] T. J. Darrell and A. P. Pentland, Recognition of Space-Time Gestures using a Distributed Representation, MIT Media Laboratory Vision and Modeling TR-197.
[8] R. O. Duda, P. E. Hart and D. G. Stork, Pattern Classification, Wiley Interscience, 2002.
[9] A. Efros, A. Berg, G. Mori and J. Malik, Recognizing Action at a Distance. ICCV, pp. 726?733,
2003.
[10] W. T. Freeman and E. H. Adelson, The Design and Use of Steerable Filters, IEEE Trans. PAMI,
13:9, pp. 891-906, 1991.
[11] T. Hastie and R. Tibshirani, Discriminant Analysis by Gaussian Mixtures. Journal of Royal
Statistical Society Series B, 58(1):155-176.
[12] W. Hawkins, P. Leichner and N. Yang, The Circular Harmonic Transform for SPECT Reconstruction and Boundary Conditions on the Fourier Transform of the Sinogram, IEEE Trans. on
Medical Imaging, 7:2, 1988.
[13] A. Heap and D. Hogg, Wormholes in Shape Space: Tracking through Discontinuous Changes
in Shape, ICCV, 1998.
[14] M. K. Hu, Visual pattern recognition by moment invariants, IEEE Trans. Inform. Theory,
8:179-187, 1962.
[15] M. J. Jones and J. M. Rehg, Statistical Color Models with Application to Skin Detection Int. J.
of Computer Vision, 46:1 pp: 81-96, 2002.
[16] B. Julesz, Textons, the elements of texture perception, and their interactions. Nature, 290:91-97,
1981.
[17] T. Leung and J. Malik, Representing and Recognizing the Visual Appearance of Materials using
Three-dimensional Textons, Int. Journal of Computer Vision, 41:1, pp. 29-44, 2001.
[18] M. Maila and J. Shi, Learning Segmentation with Random Walk, NIPS 2001.
[19] B. Moghaddam and A. Pentland, Probabilistic Visual Learning for Object Representation, IEEE
Trans. PAMI 19:7, 1997.
[20] A. Ng, M. Jordan and Y. Weiss, On Spectral Clustering: Analysis and an algorithm, NIPS,
2001.
[21] S. Satoh, Generalized Histogram: Empirical Optimization of Low Dimensional Features for
Image Matching, ECCV, 2004.
[22] J. Shi and J. Malik, Normalized Cuts and Image Segmentation, IEEE Trans. on PAMI, 2000.
[23] B. Stenger, A. Thayananthan, P. H. S. Torr, and R. Cipolla, Filtering Using a Tree-Based
Estimator, ICCV, II:1063-1070, 2003.
[24] C. Tomasi, S. Petrov and A. Sastry, 3D tracking = classification + interpolation, ICCV, 2003.
[25] R. Vidal and R. Hartley, Motion Segmentation with Missing Data using PowerFactorization
and GPCA, CVPR, 2004.
[26] Y. Weiss, Segmentation using eigenvectors: A Unifying view. ICCV, 1999.
[27] Lihi Zelnik-Manor and Michal Irani, Event-based video analysis, CVPR, 2001.
| 2640 |@word determinant:1 eliminating:1 bigram:3 proportion:1 duda:1 hu:1 zelnik:1 rgb:1 decomposition:2 covariance:1 tr:1 accommodate:1 hager:2 reduction:4 moment:4 initial:4 series:2 exclusively:1 seriously:1 bhattacharyya:2 past:1 current:1 segmentaion:1 michal:1 scatter:3 john:2 numerical:1 subsequent:1 blur:1 informative:1 shape:6 designed:1 alone:1 cue:1 selected:1 fewer:1 plane:2 gpca:3 coarse:1 successive:3 simpler:1 along:1 framewise:2 direct:1 c2:5 incorrect:1 consists:1 combine:1 interscience:1 introduce:2 pairwise:1 mask:3 expected:1 subdividing:4 multi:3 freeman:1 globally:1 relying:1 automatically:1 curse:1 window:3 considering:1 becomes:1 provided:1 project:2 bounded:1 estimating:1 mass:2 medium:1 what:1 kg:2 kind:1 kybernetika:1 differing:1 finding:1 guarantee:1 temporal:27 every:1 exactly:1 partitioning:2 medical:1 producing:1 t1:9 positive:1 aggregating:1 local:1 puppetry:1 id:4 laurent:1 interpolation:1 pami:4 initialization:4 co:1 limited:2 bi:1 unique:1 testing:3 procedure:1 steerable:1 empirical:2 jhu:2 significantly:1 projection:3 matching:4 integrating:1 bowden:1 get:1 cannot:2 operator:1 pskin:7 equivalent:1 map:1 center:7 shi:2 missing:1 straightforward:1 cluttered:1 resolution:2 estimator:1 rehg:1 stability:1 searching:2 variation:1 target:2 play:1 element:1 recognition:21 utilized:1 cut:1 database:1 binning:2 powerfactorization:1 role:1 module:1 initializing:2 solved:1 thousand:2 calculate:2 region:2 morphological:1 ordering:1 complexity:1 constrains:1 dynamic:13 trained:2 depend:1 tight:1 segment:2 easily:2 joint:1 stylized:1 represented:1 finger:1 articulated:5 train:2 separated:1 describe:3 effective:3 aggregate:1 choosing:2 refined:1 quite:1 whose:1 larger:3 solve:1 cvpr:3 otherwise:1 compressed:1 statistic:3 transform:4 noisy:1 sequence:16 advantage:2 descriptive:2 eigenvalue:3 blob:1 propose:4 subtracting:2 reconstruction:1 interaction:1 relevant:1 loop:1 multiresolution:2 achieve:3 normalize:1 convergence:1 cluster:34 darrell:2 empty:1 extending:2 converges:1 object:10 tk:3 depending:1 help:2 pose:1 exemplar:1 subregion:2 edward:1 strong:1 c:1 shadow:1 direction:1 radius:1 hartley:1 discontinuous:1 filter:1 human:2 material:1 bin:6 clustered:3 generalization:1 preliminary:1 singularity:1 hawkins:1 considered:2 trigram:2 achieves:1 efros:1 heap:1 purpose:3 estimation:3 bag:1 label:11 largest:1 create:3 weighted:1 mit:1 gaussian:9 manor:1 rather:1 pn:2 probabilistically:1 focus:1 improvement:1 properly:1 mainly:1 centroid:2 leung:1 hidden:1 comprising:1 pixel:5 issue:1 arg:3 flexible:1 among:1 classification:3 exponent:1 enrich:1 spatial:2 smoothing:8 constrained:2 integration:2 once:1 extraction:5 having:1 sampling:1 manually:1 ng:1 adelson:1 jones:1 lihi:1 foreground:6 future:1 t2:12 few:2 opening:2 recognize:6 individual:1 occlusion:1 detection:1 interest:1 circular:8 mixture:4 extreme:1 accurate:4 moghaddam:1 edge:1 orthogonal:1 tree:1 euclidean:1 taylor:1 walk:1 circle:3 re:2 theoretical:1 classify:1 modeling:10 cover:1 cost:2 recognizing:4 thacker:1 stored:1 scanning:1 gregory:1 accomplish:1 density:17 sensitivity:1 stenger:1 probabilistic:4 together:2 hopkins:2 squared:1 central:2 containing:4 choose:2 american:2 unordered:1 int:2 cremers:1 textons:2 view:5 h1:1 red:1 aggregation:1 sort:1 contribution:1 minimize:1 square:1 formed:1 accuracy:1 descriptor:4 variance:2 efficiently:1 largely:1 comparably:1 lu:1 annulus:3 trajectory:2 maxc:1 inform:1 against:1 petrov:1 energy:4 frequency:1 pp:6 obvious:1 naturally:1 proof:1 static:9 athitsos:1 sampled:2 gain:2 color:8 dimensionality:1 improves:1 segmentation:18 actually:1 higher:1 restarts:3 wei:2 rand:1 done:1 furthermore:2 xa:1 correlation:2 hand:26 lack:1 continuity:1 mode:3 stably:1 lda:13 artifact:1 building:1 effect:1 validity:1 contain:4 true:1 ranged:1 normalized:1 evolution:2 irani:1 laboratory:1 indistinguishable:1 during:1 self:2 ambiguous:1 covering:1 noted:1 cosine:2 generalized:2 pcn:1 motion:27 temperature:1 image:48 wise:2 variational:1 harmonic:1 rotation:8 stork:1 insensitive:1 million:1 extend:1 measurement:1 smoothness:3 consistency:1 hogg:1 sastry:1 closing:2 satoh:1 ssd:1 language:3 stable:2 similarity:6 surface:1 etc:1 base:1 align:1 dominant:1 aligning:1 closest:1 retrieved:1 mint:1 binary:1 ht1:3 seen:1 additional:1 employed:3 freely:1 subtraction:2 aggregated:5 recognized:1 ii:1 sliding:3 multiple:1 smooth:1 segmented:2 gesture:19 long:1 divided:1 hart:1 coded:1 impact:1 variant:1 vision:6 circumstance:1 metric:2 histogram:36 kernel:3 represent:1 normalization:2 texton:1 pyramid:3 achieved:2 c1:3 iteration:3 addition:1 residue:1 background:3 want:1 baltimore:2 annealing:1 diagram:1 singular:1 source:1 unlike:1 wormhole:1 jordan:1 yang:1 intermediate:7 independence:1 fit:3 variate:1 hastie:1 opposite:1 inner:1 handled:1 cause:1 action:40 adequate:1 useful:1 clear:1 eigenvectors:2 julesz:1 younes:2 category:5 sign:2 estimated:1 track:1 per:2 correctly:1 tibshirani:1 discrete:1 key:1 four:1 gmm:20 clean:1 undertaken:1 imaging:2 relaxation:1 sum:1 letter:1 cootes:1 place:1 scaling:5 pc2:1 spect:1 distinguish:2 comparision:1 constraint:2 precisely:1 fourier:5 speed:1 min:1 extremely:2 performing:1 expanded:1 px:1 relatively:1 department:1 belonging:1 describes:1 across:1 em:1 hl:2 intuitively:1 invariant:10 outlier:1 iccv:6 taken:3 tier:1 mori:1 equation:2 previously:1 discus:2 needed:1 vidal:1 apply:3 spectral:11 appropriate:1 alternative:1 original:1 denotes:1 clustering:36 include:2 top:1 unifying:1 pt2:1 build:7 society:1 warping:1 skin:8 intend:2 question:1 already:2 posture:4 realized:1 malik:3 md:2 spelling:1 affinity:4 subspace:6 distance:5 simulated:1 capacity:1 hmm:2 outer:1 argue:1 discriminant:2 unstable:2 reason:1 length:1 modeled:1 tiered:5 ratio:1 difficult:3 negative:1 design:1 proper:2 adjustable:2 perform:2 upper:1 observation:4 markov:1 pentland:3 optional:1 extended:1 frame:26 locate:1 varied:1 pc1:1 community:1 concentric:1 required:1 c3:1 extensive:1 tomasi:2 learned:3 nip:2 trans:6 beyond:1 pattern:3 perception:1 including:3 reliable:1 video:5 royal:1 power:3 critical:2 event:1 natural:1 rely:1 regularized:1 representing:2 improve:1 library:4 temporally:3 dtw:1 hm:1 extract:3 kf:2 determining:1 filtering:1 h2:1 integrate:1 sufficient:2 consistent:1 viewpoint:3 classifying:1 translation:2 eccv:1 summary:1 changed:1 bias:1 template:1 absolute:2 distributed:1 boundary:1 dimension:8 calculated:1 gram:2 transition:3 contour:2 refinement:3 projected:1 adaptive:6 compact:1 global:2 active:1 incoming:1 discriminative:3 spectrum:2 subsequence:1 iterative:1 decade:1 learn:2 nature:1 complex:1 domain:1 pk:1 body:5 representative:2 fashion:1 wiley:1 sub:1 position:1 candidate:1 third:1 weighting:3 unigram:1 list:1 incorporating:3 thayananthan:1 quantization:1 sequential:2 ci:1 texture:2 magnitude:1 accomodate:1 labelling:2 sclaroff:1 subtract:1 simply:1 appearance:15 visual:4 ordered:1 tracking:4 cipolla:1 extracted:2 slot:1 goal:1 presentation:2 kmeans:1 change:6 determined:3 torr:1 uniformly:1 principal:2 invariance:1 ya:2 subdivision:1 brand:1 htm1:1 berg:1 scan:1 violated:1 evaluate:1 tested:1 |
1,807 | 2,641 | Incremental Learning for Visual Tracking
?
Jongwoo Lim? David Ross? Ruei-Sung Lin? Ming-Hsuan Yang?
University of Illinois ? University of Toronto ? Honda Research Institute
[email protected] [email protected] [email protected] [email protected]
Abstract
Most existing tracking algorithms construct a representation of a target
object prior to the tracking task starts, and utilize invariant features to
handle appearance variation of the target caused by lighting, pose, and
view angle change. In this paper, we present an efficient and effective online algorithm that incrementally learns and adapts a low dimensional eigenspace representation to reflect appearance changes of the target, thereby facilitating the tracking task. Furthermore, our incremental
method correctly updates the sample mean and the eigenbasis, whereas
existing incremental subspace update methods ignore the fact the sample
mean varies over time. The tracking problem is formulated as a state
inference problem within a Markov Chain Monte Carlo framework and
a particle filter is incorporated for propagating sample distributions over
time. Numerous experiments demonstrate the effectiveness of the proposed tracking algorithm in indoor and outdoor environments where the
target objects undergo large pose and lighting changes.
1
Introduction
The main challenges of visual tracking can be attributed to the difficulty in handling appearance variability of a target object. Intrinsic appearance variabilities include pose variation
and shape deformation of a target object, whereas extrinsic illumination change, camera
motion, camera viewpoint, and occlusions inevitably cause large appearance variation. Due
to the nature of the tracking problem, it is imperative for a tracking algorithm to model such
appearance variation.
Here we developed a method that, during visual tracking, constantly and efficiently updates a low dimensional eigenspace representation of the appearance of the target object.
The advantages of this adaptive subspace representation are several folds. The eigenspace
representation provides a compact notion of the ?thing? being tracked rather than treating
the target as a set of independent pixels, i.e., ?stuff? [1]. The use of an incremental method
continually updates the eigenspace to reflect the appearance change caused by intrinsic
and extrinsic factors, thereby facilitating the tracking process. To estimate the locations
of the target objects in consecutive frames, we used a sampling algorithm with likelihood
estimates, which is in direct contrast to other tracking methods that usually solve complex
optimization problems using gradient-descent approach.
The proposed method differs from our prior work [14] in several aspects. First, the proposed algorithm does not require any training images of the target object before the tracking
task starts. That is, our tracker learns a low dimensional eigenspace representation on-line
and incrementally updates it as time progresses (We assume, like most tracking algorithms,
that the target region has been initialized in the first frame). Second, we extend our sampling method to incorporate a particle filter so that the sample distributions are propagated
over time. Based on the eigenspace model with updates, an effective likelihood estimation
function is developed. Third, we extend the R-SVD algorithm [6] so that both the sample
mean and eigenbasis are correctly updated as new data arrive. Though there are numerous
subspace update algorithms in the literature, only the method by Hall et al. [8] is also able
to update the sample mean. However, their method is based on the addition of a single column (single observation) rather than blocks (a number of observations in our case) and thus
is less efficient than ours. While our formulation provides an exact solution, their algorithm
gives only approximate updates and thus it may suffer from numerical instability. Finally,
the proposed tracker is extended to use a robust error norm for likelihood estimation in the
presence of noisy data or partial occlusions, thereby rendering more accurate and robust
tracking results.
2
Previous Work and Motivation
Black et al. [4] proposed a tracking algorithm using a pre-trained view-based eigenbasis
representation and a robust error norm. Instead of relying on the popular brightness constancy working principal, they advocated the use of subspace constancy assumption for
visual tracking. Although their algorithm demonstrated excellent empirical results, it requires to build a set of view-based eigenbases before the tracking task starts. Furthermore,
their method assumes that certain factors, such as illumination conditions, do not change
significantly as the eigenbasis, once constructed, is not updated.
Hager and Belhumeur [7] presented a tracking algorithm to handle the geometry and illumination variations of target objects. Their method extends a gradient-based optical flow
algorithm to incorporate research findings in [2] for object tracking under varying illumination conditions. Prior to the tracking task starts, a set of illumination basis needs to be
constructed at a fixed pose in order to account for appearance variation of the target due to
lighting changes. Consequently, it is not clear whether this method is effective if a target
object undergoes changes in illumination with arbitrary pose.
In [9] Isard and Blake developed the Condensation algorithm for contour tracking in which
multiple plausible interpretations are propagated over time. Though their probabilistic approach has demonstrated success in tracking contours in clutter, the representation scheme
is rather primitive, i.e., curves or splines, and is not updated as the appearance of a target
varies due to pose or illumination change.
Mixture models have been used to describe appearance change for motion estimation [3]
[10]. In Black et al. [3] four possible causes are identified in a mixture model for estimating
appearance change in consecutive frames, and thereby more reliable image motion can be
obtained. A more elaborate mixture model with an online EM algorithm was recently
proposed by Jepson et al. [10] in which they use three components and wavelet filters to
account for appearance changes during tracking. Their method is able to handle variations
in pose, illumination and expression. However, their WSL appearance model treats pixels
within the target region independently, and therefore does not have notion of the ?thing?
being tracked. This may result in modeling background rather than the foreground, and fail
to track the target.
In contrast to the eigentracking algorithm [4], our algorithm does not require a training
phase but learns the eigenbases on-line during the object tracking process, and constantly
updates this representation as the appearance changes due to pose, view angle, and illumination variation. Further, our method uses a particle filter for motion parameter estimation
rather than the Gauss-Newton method which often gets stuck in local minima or is distracted by outliers [4]. Our appearance-based model provides a richer description than
simple curves or splines as used in [9], and has notion of the ?thing? being tracked. In
addition, the learned representation can be utilized for other tasks such as object recognition. In this work, an eigenspace representation is learned directly from pixel values
within a target object in the image space. Experiments show that good tracking results
can be obtained with this representation without resorting to wavelets as used in [10], and
better performance can potentially be achieved using wavelet filters. Note also that the
view-based eigenspace representation has demonstrated its ability to model appearance of
objects at different pose [13], and under different lighting conditions [2].
3
Incremental Learning for Tracking
We present the details of the proposed incremental learning algorithm for object tracking
in this section.
3.1 Incremental Update of Eigenbasis and Mean
The appearance of a target object may change drastically due to intrinsic and extrinsic
factors as discussed earlier. Therefore it is important to develop an efficient algorithm to
update the eigenspace as the tracking task progresses. Numerous algorithms have been
developed to update eigenbasis from a time-varying covariance matrix as more data arrive
[6] [8] [11] [5]. However, most methods assume zero mean in updating the eigenbasis
except the method by Hall et al. [8] in which they consider the change of the mean when
updating eigenbasis as each new datum arrives. Their update algorithm only handles one
datum per update and gives approximate results, while our formulation handles multiple
data at the same time and renders exact solutions.
We extend the work of the classic R-SVD method [6] in which we update the eigenbasis
while taking the shift of the sample mean into account. To the best of our knowledge, this
formulation with mean update is new in the literature.
Given a d ? n data matrix A = {I1 , . . . , In } where each column Ii is an observation (a ddimensional image vector in this paper), we can compute the singular value decomposition
(SVD) of A, i.e., A = U ?V >. When a d?m matrix E of new observations is available, the
0
0
0
0
R-SVD algorithm efficiently computes the SVD of the matrix A = (A|E) = U ? V >
based on the SVD of A as follows:
? of E, and U 0 = (U |E).
?
1. Apply QR decomposition to and get orthonormal basis E
0
2. Let V = V0 I0m where Im is an m ? m identity matrix. It follows then,
>
U >AV U >E ? U >E
0
0
0
0
V 0
U
? = U >A V = E
= 0 E? >E .
? > (A|E) 0 Im = E
? >AV E
? >E
0
??
? V? > and the SVD of A0 is
3. Compute the SVD of ? = U
0
0
??
? V? >)V 0 > = (U 0 U
? )?(
? V? >V 0 > ).
A = U (U
Exploiting the properties of orthonormal bases and block structures, the R-SVD algorithm
computes the new eigenbasis efficiently. The computational complexity analysis and more
details are described in [6].
One problem with the R-SVD algorithm is that the eigenbasis U is computed from AA >
with the zero mean assumption. We modify the R-SVD algorithm and compute the eigenbasis with mean update. The following derivation is based on scatter matrix, which is same
as covariance matrix except a scalar factor.
Proposition 1 Let Ip = {I1 , I2 , . . . , In }, Iq = {In+1 , In+2 , . . . , In+m }, and Ir = (Ip |Iq ).
Denote the means and scatter matrices of Ip , Iq , Ir as ?Ip , ?Iq , ?Ir , and Sp , Sq , Sr respecnm ?
tively, then Sr = Sp + Sq + n+m
(Iq ? ?Ip )(?Iq ? ?Ip )> .
n ?
m ?
m ?
Ip + n+m
Iq , ?Ip ? ?Ir = n+m
Proof: By definition, ?Ir = n+m
(Ip ? ?Iq ); ?Iq ? ?Ir =
n
?
?
n+m (Iq ? Ip ) and,
Pn
Pn+m
Sr = Pi=1 (Ii ? ?Ir )(Ii ? ?Ir )> + i=n+1 (Ii ? ?Ir )(Ii ? ?Ir )>
n
=
(Ii ? ?Ip + ?Ip ? ?Ir )(Ii ? ?Ip + ?Ip ? ?Ir )> +
Pi=1
n+m
?
?
?
?
?
? >
i=m+1 (Ii ? Iq + Iq ? Ir )(Ii ? Iq + Iq ? Ir )
= Sp + n(?Ip ? ?Ir )(?Ip ? ?Ir )> + Sq + m(?Iq ? ?Ir )(?Iq ? ?Ir )>
nm2
?
? ?
? > + Sq + n2 m 2 (?Ip ? ?Iq )(?Ip ? ?Iq )>
= Sp + (n+m)
2 (Ip ? Iq )(Ip ? Iq )
(n+m)
nm ?
= Sp + Sq + n+m
(Ip ? ?Iq )(?Ip ? ?Iq )>
? and I?r = {I1 ?
Let I?p = {I1 ? ?Ip , . . . , In ? ?Ip }, I?q = {In+1 ? ?Iq , . . . , In+m ?
Iq },
q
?Ir , . . . , In+m ? ?Ir }, and the SVD of I?r = Ur ?r V > . Let E
? = I?q | nm (?Ip ? ?Iq ) ,
r
n+m
? I?p |E)
? >. Therefore, we compute SVD on (I?p |E)
? to
and use Proposition 1, Sr = (I?p |E)(
get Ur . This can be done efficiently by the R-SVD algorithm as described above.
In summary, given the mean ?Ip and the SVD of existing data Ip , i.e., Up ?p Vp> and new
data Iq , we can compute the the mean ?Ir and the SVD of Ir , i.e., Ur ?r Vr> easily:
q
n ?
m ?
? = Iq ? ?Ir 1(1?m) | nm (?Ip ? ?Iq ) .
1. Compute ?Ir = n+m
Ip + n+m
Iq , and E
n+m
? to obtain (Ur ?r Vr>).
2. Compute R-SVD with (Up ?p Vp>) and E
In numerous vision problems, we can further exploit the low dimensional approximation of
image data and put larger weights on the recent observations, or equivalently downweight
the contributions of previous observations. For example as the appearance of a target object
gradually changes, we may want to put more weights on recent observations in updating
the eigenbasis since they are more likely to be similar to the current appearance of the
target. The forgetting factor f can be used under this premise as suggested in [11] , i.e.,
0
0
A = (f A |E) = (U (f ?)V >|E) where A and A are original and weighted data matrices,
respectively.
3.2 Sequential Inference Model
The visual tracking problem is cast as an inference problem with a Markov model and
hidden state variable, where a state variable Xt describes the affine motion parameters
(and thereby the location) of the target at time t. Given a set of observed images I t =
{I1 , . . . , It }. we aim to estimate the value of the hidden state variable Xt . Using Bayes?
theorem, we have
Z
p(Xt | It ) ? p(It |Xt )
p(Xt |Xt?1 ) p(Xt?1 | It?1 ) dXt?1
The tracking process is governed by the observation model p(It |Xt ) where we estimate the
likelihood of Xt observing It , and the dynamical model between two states p(Xt |Xt?1 ).
The Condensation algorithm [9], based on factored sampling, approximates an arbitrary
distribution of observations with a stochastically generated set of weighted samples. We
use a variant of the Condensation algorithm to model the distribution over the object?s
location, as it evolves over time.
3.3 Dynamical and Observation Models
The motion of a target object between two consecutive frames can be approximated by
an affine image warping. In this work, we use the six parameters of affine transform
to model the state transition from Xt?1 to Xt of a target object being tracked. Let
Xt = (xt , yt , ?t , st , ?t , ?t ) where xt , yt , ?t , st , ?t , ?t , denote x, y translation, rotation
angle, scale, aspect ratio, and skew direction at time t. Each parameter in Xt is modeled
independently by a Gaussian distribution around its counterpart in Xt?1 . That is,
p(Xt |Xt?1 ) = N (Xt ; Xt?1 , ?)
where ? is a diagonal covariance matrix whose elements are the corresponding variances
of affine parameters, i.e., ?x2 , ?y2 , ??2 , ?s2 , ??2 , ??2 .
Since our goal is to use a representation to model the ?thing? that we are tracking, we
model the image observations using a probabilistic interpretation of principal component
analysis [16]. Given an image patch predicated by Xt , we assume the observed image It
was generated from a subspace spanned by U centered at ?. The probability that a sample
being generated from the subspace is inversely proportional to the distance d from the
sample to the reference point (i.e., center) of the subspace, which can be decomposed into
the distance-to-subspace, dt , and the distance-within-subspace from the projected sample
to the subspace center, dw . This distance formulation, based on a orthonormal subspace
and its complement space, is similar to [12] in spirit.
The probability of a sample generated from a subspace, pdt (It |Xt ), is governed by a Gaussian distribution:
pdt (It | Xt ) = N (It ; ?, U U >+ ?I)
where I is an identity matrix, ? is the mean, and ?I term corresponds to the additive Gaussian noise in the observation process. It can be shown [15] that the negative exponential
distance from It to the subspace spanned by U , i.e., exp(?||(It ? ?) ? U U >(It ? ?)||2 ),
is proportional to N (It ; ?, U U >+ ?I) as ? ? 0.
Within a subspace, the likelihood of the projected sample can be modeled by the Mahalanobis distance from the mean as follows:
pdw (It | Xt ) = N (It ; ?, U ??2 U >)
where ? is the center of the subspace and ? is the matrix of singular values corresponding
to the columns of U . Put together, the likelihood of a sample being generated from the
subspace is governed by
p(It |Xt ) = pdt (It |Xt ) pdw (It |Xt ) = N (It ; ?, U U > + ?I) N (It ; ?, U ??2 U > ) (1)
Given a drawn sample Xt and the corresponding image region It , we aim to compute
p(It |Xt ) using (1). To minimize the effects of noisy pixels, we utilize a robust error norm
2
[4], ?(x, ?) = ?2x+x2 instead of the Euclidean norm d(x) = ||x||2 , to ignore the ?outlier?
pixels (i.e., the pixels that are not likely to appear inside the target region given the current
eigenspace). We use a method similar to that used in [4] in order to compute dt and dw .
This robust error norm is helpful especially when we use a rectangular region to enclose
the target (which inevitably contains some noisy background pixels).
4
Experiments
To test the performance of our proposed tracker, we collected a number of videos recorded
in indoor and outdoor environments where the targets change pose in different lighting conditions. Each video consists of 320 ? 240 gray scale images and are recorded at 15 frames
per second unless specified otherwise. For the eigenspace representation, each target image
region is resized to 32 ? 32 patch, and the number of eigenvectors used in all experiments
is set to 16 though fewer eigenvectors may also work well. Implemented in MATLAB
with MEX, our algorithm runs at 4 frames per second on a standard computer with 200
particles. We present some tracking results in this section and more tracking results as well
as videos can be found at http://vision.ucsd.edu/?jwlim/ilt/.
4.1 Experimental Results
Figure 1 shows the tracking results using a challenging sequence recorded with a moving digital camera in which a person moves from a dark room toward a bright area while
changing his pose, moving underneath spot lights, changing facial expressions and taking
off glasses. All the eigenbases are constructed automatically from scratch and constantly
updated to model the appearance of the target object while undergoing appearance changes.
Even with the significant camera motion and low frame rate (which makes the motions between frames more significant, or equivalently to tracking fast moving objects), our tracker
stays stably on the target throughout the sequence.
The second sequence contains an animal doll moving in different pose, scale, and lighting
conditions as shown in Figure 2. Experimental results demonstrate that our tracker is able
to follow the target as it undergoes large pose change, cluttered background, and lighting
variation. Notice that the non-convex target object is localized with an enclosing rectangular window, and thus it inevitably contains some background pixels in its appearance
representation. The robust error norm enables the tracker to ignore background pixels and
estimate the target location correctly. The results also show that our algorithm faithfully
Figure 1: A person moves from dark toward bright area with large lighting and pose changes. The
images in the second row shows the current sample mean, tracked region, reconstructed image, and
the reconstruction error respectively. The third and forth rows shows 10 largest eigenbases.
Figure 2: An animal doll moving with large pose, lighting variation in a cluttered background.
models the appearance of the target, as shown in eigenbases and reconstructed images, in
the presence of noisy background pixels.
We recorded a sequence to demonstrate that our tracker performs well in outdoor environment where lighting conditions change drastically. The video was acquired when a person
walking underneath a trellis covered by vines. As shown in Figure 3, the cast shadow
changes the appearance of the target face drastically. Furthermore, the combined pose and
lighting variation with low frame rate makes the tracking task extremely difficult. Nevertheless, the results show that our tracker successfully follows the target accurately and
robustly. Due to heavy shadows and drastic lighting change, other tracking methods based
on gradient, contour, or color information are unlikely to perform well in this case.
4.2 Discussion
The success of our tracker can be attributed to several factors. It is well known that
the appearance of an object undergoing pose change can be modeled well by view-based
Figure 3: A person moves underneath a trellis with large illumination change and cast shadows
while changing his pose. More results can be found in the project web page.
representation [13]. Meanwhile at fixed pose, the appearance of an object in different
illumination conditions can be approximated well by a low dimensional subspace [2]. Our
empirical results show that these variations can be learned on-line without any prior training
phase, and also the changes caused by cast and attached shadows can still be approximated
by a linear subspace to some extent.
We show a few failure cases at our the web site
mentioned earlier. Typically, the failure happens when there is a combination of fast pose
change and drastic illumination change.
In this paper, we do not directly address the partial occlusion problems. Empirical results
show that temporary and partial occlusions can be handled by our method through the
robust error norm and the constant update of the eigenspace. Nevertheless situations arise
where we may have prior knowledge of the objects being tracked, and can exploit such
information for better occlusion handling.
To demonstrate the potency of our modified R-SVD algorithm in faithfully modeling the
object appearance, we compare the reconstructed images using our method and a conventional SVD algorithm. In Figure 4 first row contains a set of images tracked by our
tracker, and the second and fourth rows show the reconstructed images using 16 eigenvectors obtained after 121 incremental updates of 605 frame (block size is set to 5), and the
top 16 eigenvectors obtained by conventional SVD algorithm using all 605 tracked images.
Note that we only maintained 16 eigenvectors during tracking, and discarded the remaining
eigenvectors at each update. The residue images are presented in the third and fifth rows,
and the average L2 reconstruction error per pixel is 5.73?10?2 and 5.65?10?2 for our
modified R-SVD method and the conventional SVD algorithm respectively. The figure and
average reconstruction error shows that our modified R-SVD method is able to effectively
model the object appearance without losing detailed information.
5
Conclusions and Future Work
We have presented an appearance-based tracker that incrementally learns a low dimensional
eigenspace representation for object tracking while the target undergoes pose, illumination
and appearance changes. Whereas most tracking algorithms operate on the premise that
the object appearance or ambient environment lighting condition does not change as time
progresses, our method adapts the model representation to reflect appearance variation of
the target, thereby facilitating the tracking task. In contrast to the existing incremental
subspace methods, our R-SVD method updates the mean and eigenbasis accurately and
efficiently, and thereby learns a good eigenspace representation to faithfully model the
appearance of the target being tracked. Our experiments demonstrate the effectiveness of
the proposed tracker in indoor and outdoor environments where the target objects undergo
large pose and lighting changes.
The current dynamical model in our sampling method is based on a Gaussian distribution,
but the dynamics could be learned from exemplars for more efficient parameter estimation.
Our algorithm can be extended to construct a set of eigenbases for modeling nonlinear
aspects of appearance variation more precisely and automatically. We aim to address these
issues in our future work.
Figure 4: Reconstructed images and errors using our and the conventional SVD algorithms.
References
[1] E. H. Adelson and J. R. Bergen. The plenoptic function and the elements of early vision. In M. Landy and
J. A. Movshon, editors, Computational Models of Visual Processing, pp. 1?20. MIT Press, 1991.
[2] P. Belhumeur and D. Kreigman. What is the set of images of an object under all possible lighting conditions.
In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, pp. 270?277, 1997.
[3] M. J. Black, D. J. Fleet, and Y. Yacoob. A framework for modeling appearance change in image sequence.
In Proceedings of the Sixth IEEE International Conference on Computer Vision, pp. 660?667, 1998.
[4] M. J. Black and A. D. Jepson. Eigentracking: Robust matching and tracking of articulated objects using
view-based representation. In Proceedings of European Conference on Computer Vision, pp. 329?342,
1996.
[5] M. Brand. Incremental singular value decomposition of uncertain data with missing values. In Proceedings
of the Seventh European Conference on Computer Vision, volume 4, pp. 707?720, 2002.
[6] G. H. Golub and C. F. Van Loan. Matrix Computations. The Johns Hopkins University Press, 1996.
[7] G. Hager and P. Belhumeur. Real-time tracking of image regions with changes in geometry and illumination. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, pp. 403?410,
1996.
[8] P. Hall, D. Marshall, and R. Martin. Incremental eigenanalysis for classification. In Proceedings of British
Machine Vision Conference, pp. 286?295, 1998.
[9] M. Isard and A. Blake. Contour tracking by stochastic propagation of conditional density. In Proceedings
of the Fourth European Conference on Computer Vision, volume 2, pp. 343?356, 1996.
[10] A. D. Jepson, D. J. Fleet, and T. F. El-Maraghi. Robust online appearance models for visual tracking. In
Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, volume 1, pp. 415?422,
2001.
[11] A. Levy and M. Lindenbaum. Sequential Karhunen-Loeve basis extraction and its application to images.
IEEE Transactions on Image Processing, 9(8):1371?1374, 2000.
[12] B. Moghaddam and A. Pentland. Probabilistic visual learning for object recognition. IEEE Transactions
on Pattern Analysis and Machine Intelligence, 19(7):696?710, 1997.
[13] H. Murase and S. Nayar. Visual learning and recognition of 3d objects from appearance. International
Journal of Computer Vision, 14(1):5?24, 1995.
[14] D. Ross, J. Lim, and M.-H. Yang. Adaptive probabilistic visual tracking with incremental subspace update.
In Proceedings of the Eighth European Conference on Computer Vision, volume 2, pp. 470?482, 2004.
[15] S. Roweis. EM algorithms for PCA and SPCA. In Advances in Neural Information Processing Systems
10, pp. 626?632, 1997.
[16] M. E. Tipping and C. M. Bishop. Probabilistic principal component analysis. Journal of the Royal Statistical Society, Series B, 61(3):611?622, 1999.
| 2641 |@word norm:7 covariance:3 decomposition:3 brightness:1 thereby:7 hager:2 contains:4 series:1 ours:1 existing:4 current:4 com:1 scatter:2 john:1 additive:1 numerical:1 eigentracking:2 shape:1 enables:1 treating:1 update:23 isard:2 fewer:1 intelligence:1 pdw:2 provides:3 honda:2 toronto:2 location:4 constructed:3 direct:1 consists:1 inside:1 acquired:1 forgetting:1 uiuc:2 ming:1 relying:1 decomposed:1 automatically:2 window:1 project:1 estimating:1 eigenspace:14 what:1 developed:4 finding:1 sung:1 stuff:1 appear:1 continually:1 before:2 local:1 treat:1 modify:1 dross:1 black:4 challenging:1 camera:4 block:3 differs:1 spot:1 sq:5 area:2 empirical:3 significantly:1 matching:1 pre:1 get:3 lindenbaum:1 put:3 instability:1 conventional:4 demonstrated:3 yt:2 center:3 missing:1 primitive:1 independently:2 cluttered:2 rectangular:2 convex:1 hsuan:1 factored:1 orthonormal:3 spanned:2 his:2 dw:2 classic:1 handle:5 notion:3 variation:14 updated:4 target:40 exact:2 losing:1 us:1 element:2 recognition:6 approximated:3 utilized:1 updating:3 walking:1 observed:2 constancy:2 vine:1 region:8 mentioned:1 environment:5 complexity:1 dynamic:1 trained:1 basis:3 easily:1 derivation:1 articulated:1 fast:2 effective:3 describe:1 monte:1 whose:1 richer:1 larger:1 solve:1 plausible:1 otherwise:1 ability:1 transform:1 rlin1:1 noisy:4 ip:29 online:3 advantage:1 sequence:5 reconstruction:3 adapts:2 roweis:1 forth:1 description:1 qr:1 eigenbasis:14 exploiting:1 incremental:12 object:35 iq:29 develop:1 pose:22 propagating:1 exemplar:1 advocated:1 progress:3 ddimensional:1 c:1 enclose:1 implemented:1 shadow:4 murase:1 direction:1 filter:5 stochastic:1 centered:1 require:2 premise:2 maraghi:1 proposition:2 im:2 tracker:12 hall:3 blake:2 around:1 exp:1 consecutive:3 early:1 estimation:5 ross:2 largest:1 faithfully:3 successfully:1 weighted:2 mit:1 gaussian:4 aim:3 modified:3 rather:5 pn:2 resized:1 varying:2 yacoob:1 likelihood:6 contrast:3 underneath:3 glass:1 helpful:1 inference:3 bergen:1 el:1 unlikely:1 typically:1 a0:1 hidden:2 i1:5 pixel:11 issue:1 classification:1 animal:2 construct:2 once:1 extraction:1 sampling:4 adelson:1 foreground:1 future:2 spline:2 few:1 geometry:2 occlusion:5 phase:2 golub:1 mixture:3 arrives:1 light:1 chain:1 accurate:1 ambient:1 moghaddam:1 partial:3 facial:1 unless:1 euclidean:1 initialized:1 deformation:1 uncertain:1 column:3 modeling:4 earlier:2 marshall:1 imperative:1 seventh:1 varies:2 combined:1 st:2 person:4 international:2 density:1 stay:1 probabilistic:5 off:1 together:1 hopkins:1 reflect:3 nm:3 recorded:4 stochastically:1 account:3 caused:3 view:7 observing:1 start:4 bayes:1 contribution:1 minimize:1 bright:2 ir:24 variance:1 efficiently:5 vp:2 accurately:2 carlo:1 lighting:15 definition:1 sixth:1 failure:2 pp:11 proof:1 attributed:2 propagated:2 popular:1 lim:2 knowledge:2 color:1 dt:2 tipping:1 follow:1 formulation:4 done:1 though:3 furthermore:3 predicated:1 working:1 web:2 nonlinear:1 propagation:1 incrementally:3 undergoes:3 stably:1 gray:1 effect:1 y2:1 counterpart:1 i2:1 mahalanobis:1 during:4 maintained:1 demonstrate:5 performs:1 motion:8 image:27 recently:1 rotation:1 tracked:9 tively:1 attached:1 volume:4 extend:3 interpretation:2 approximates:1 discussed:1 significant:2 resorting:1 particle:4 illinois:1 moving:5 v0:1 base:1 recent:2 certain:1 success:2 minimum:1 belhumeur:3 ii:9 multiple:2 lin:1 myang:1 variant:1 vision:13 mex:1 achieved:1 whereas:3 addition:2 condensation:3 background:7 want:1 residue:1 singular:3 operate:1 sr:4 undergo:2 thing:4 flow:1 spirit:1 effectiveness:2 yang:2 presence:2 spca:1 rendering:1 identified:1 shift:1 fleet:2 whether:1 expression:2 six:1 handled:1 pca:1 movshon:1 suffer:1 render:1 cause:2 matlab:1 clear:1 eigenvectors:6 covered:1 detailed:1 clutter:1 dark:2 http:1 notice:1 extrinsic:3 correctly:3 track:1 per:4 four:1 nevertheless:2 wsl:1 drawn:1 changing:3 utilize:2 jongwoo:1 run:1 angle:3 fourth:2 ilt:1 arrive:2 extends:1 throughout:1 patch:2 datum:2 i0m:1 fold:1 precisely:1 ri:1 x2:2 loeve:1 aspect:3 extremely:1 optical:1 downweight:1 martin:1 combination:1 describes:1 em:2 ur:4 evolves:1 happens:1 outlier:2 invariant:1 gradually:1 handling:2 skew:1 fail:1 drastic:2 available:1 doll:2 apply:1 pdt:3 robustly:1 original:1 assumes:1 top:1 include:1 remaining:1 newton:1 landy:1 exploit:2 build:1 especially:1 society:1 warping:1 move:3 ruei:1 diagonal:1 gradient:3 subspace:20 distance:6 collected:1 extent:1 toward:2 modeled:3 ratio:1 equivalently:2 difficult:1 potentially:1 negative:1 enclosing:1 perform:1 av:2 observation:12 markov:2 discarded:1 descent:1 inevitably:3 pentland:1 situation:1 extended:2 incorporated:1 variability:2 frame:10 distracted:1 ucsd:1 arbitrary:2 david:1 complement:1 cast:4 specified:1 learned:4 nm2:1 temporary:1 address:2 able:4 suggested:1 usually:1 dynamical:3 pattern:4 indoor:3 eighth:1 challenge:1 reliable:1 royal:1 video:4 difficulty:1 scheme:1 inversely:1 numerous:4 prior:5 literature:2 l2:1 dxt:1 proportional:2 localized:1 digital:1 affine:4 viewpoint:1 editor:1 pi:2 heavy:1 translation:1 row:5 summary:1 drastically:3 institute:1 taking:2 face:1 fifth:1 van:1 curve:2 transition:1 contour:4 computes:2 stuck:1 adaptive:2 projected:2 transaction:2 reconstructed:5 approximate:2 compact:1 ignore:3 nature:1 robust:9 excellent:1 complex:1 meanwhile:1 european:4 jepson:3 sp:5 main:1 motivation:1 s2:1 noise:1 arise:1 n2:1 facilitating:3 site:1 elaborate:1 vr:2 trellis:2 exponential:1 outdoor:4 governed:3 levy:1 third:3 learns:5 wavelet:3 theorem:1 british:1 xt:31 bishop:1 undergoing:2 eigenanalysis:1 intrinsic:3 sequential:2 effectively:1 illumination:14 karhunen:1 appearance:38 likely:2 visual:10 tracking:47 scalar:1 aa:1 corresponds:1 constantly:3 conditional:1 identity:2 formulated:1 goal:1 consequently:1 room:1 change:33 loan:1 except:2 principal:3 svd:25 gauss:1 experimental:2 brand:1 plenoptic:1 incorporate:2 scratch:1 nayar:1 |
1,808 | 2,642 | Adaptive Discriminative Generative Model
and Its Applications
?
Ruei-Sung Lin? David Ross? Jongwoo Lim? Ming-Hsuan Yang?
University of Illinois ? University of Toronto ? Honda Research Institute
[email protected] [email protected] [email protected] [email protected]
Abstract
This paper presents an adaptive discriminative generative model that generalizes the conventional Fisher Linear Discriminant algorithm and renders a proper probabilistic interpretation. Within the context of object
tracking, we aim to find a discriminative generative model that best separates the target from the background. We present a computationally
efficient algorithm to constantly update this discriminative model as time
progresses. While most tracking algorithms operate on the premise that
the object appearance or ambient lighting condition does not significantly
change as time progresses, our method adapts a discriminative generative model to reflect appearance variation of the target and background,
thereby facilitating the tracking task in ever-changing environments. Numerous experiments show that our method is able to learn a discriminative generative model for tracking target objects undergoing large pose
and lighting changes.
1 Introduction
Tracking moving objects is an important and essential component of visual perception,
and has been an active research topic in computer vision community for decades. Object
tracking can be formulated as a continuous state estimation problem where the unobservable states encode the locations or motion parameters of the target objects, and the task is
to infer the unobservable states from the observed images over time. At each time step,
a tracker first predicts a few possible locations (i.e., hypotheses) of the target in the next
frame based on its prior and current knowledge. The prior knowledge includes its previous
observations and estimated state transitions. Among these possible locations, the tracker
then determines the most likely location of the target object based on the new observation. An attractive and effective prediction mechanism is based on Monte Carlo sampling
in which the state dynamics (i.e., transition) can be learned with a Kalman filter or simply
modeled as a Gaussian distribution. Such a formulation indicates that the performance of
a tracker is largely based on a good observation model for validating all hypotheses. Indeed, learning a robust observation model has been the focus of most recent object tracking
research within this framework, and is also the focus of this paper.
Most of the existing approaches utilize static observation models and construct them before
a tracking task starts. To account for all possible variation in a static observation model,
it is imperative to collect a large set of training examples with the hope that it covers all
possible variations of the object?s appearance. However, it is well known that the appearance of an object varies significantly under different illumination, viewing angle, and shape
deformation. It is a daunting, if not impossible, task to collect a training set that enumerates
all possible cases. An alternative approach is to develop an adaptive method that contains
a number of trackers that track different features or parts of a target object [3]. Therefore,
even though each tracker may fail under certain circumstances, it is unlikely all of them fail
at the same time. The tracking method then adaptively selects the trackers that are robust at
current situation to predict object locations. Although this approach improves the flexibility
and robustness of a tracking method, each tracker has a static observation model which has
to be trained beforehand and consequently restricts its application domains severely. There
are numerous cases, e.g., robotics applications, where the tracker is expected to track a previously unseen target once it is detected. To the best of our knowledge, considerably less
attention is paid to developing adaptive observation models to account for appearance variation of a target object (e.g., pose, deformation) or environmental changes (e.g., lighting
conditions and viewing angles) as tracking task progresses.
Our approach is to learn a model for determining the probability of a predicted image location being generated from the class of the target or the background. That is, we formulate
a binary classification problem and develop a discriminative model to distinguish observations from the target class and the background class. While conventional discriminative
classifiers simply predict the class of each test sample, a good model within the abovementioned tracking framework needs to select the most likely sample that belongs to target
object class from a set of samples (or hypotheses). In other words, an observation model
needs a classifier with proper probabilistic interpretation.
In this paper, we present an adaptive discriminative generative model and apply it to object
tracking. The proposed model aims to best separate the target and the background in the
ever-changing environment. The problem is formulated as a density estimation problem,
where the goal is, given a set of positive (i.e., belonging to the target object class) and negative examples (i.e., belonging to the background class), to learn a distribution that assigns
high probability to the positive examples and low probability to the negative examples. This
is done in a two-stage process. First, in the generative stage, we use a probabilistic principal
component analysis to model the density of the positive examples. The result of this state is
a Gaussian, which assigns high probability to examples lying in the linear subspace which
captures the most variance of the positive examples. Second, in the discriminative stage,
we use negative examples (specifically, negative examples that are assigned high probability by our generative model) to produce a new distribution which reduces the probability
of the negative examples. This is done by learning a linear projection that, when applied
to the data and the generative model, increases the distance between the negative examples
and the mean. Toward that end, it is formulated as an optimization problem and we show
that this is a direct generalization of the conventional Fisher Linear Discriminant algorithm
with proper probabilistic interpretation. Our experimental results show that our algorithm
can reliably track moving objects whose appearance changes under different poses, illumination, and self deformation.
2 Probabilistic Tracking Algorithm
We formulate the object tracking problem as a state estimation problem in a way similar
to [5] [9]. Denote ot as an image region observed at time t and Ot = {o1 , . . . , ot } is a set
of image regions observed from the beginning to time t. An object tracking problem is a
process to infer state st from observation Ot , where state st contains a set of parameters
referring to the tracked object?s 2-D position, orientation, and scale in image ot . Assuming
a Markovian state transition, this inference
Z problem is formulated with a recursive equation:
p(st |Ot ) = kp(ot |st ) p(st |st?1 )p(st?1 |Ot?1 )dst?1
(1)
where k is a constant, and p(ot |st ) and p(st |st?1 ) correspond to the observation model and
dynamic model, respectively.
In (1), p(st?1 |Ot?1 ) is the state estimation given all the prior observations up to time t ? 1,
and p(ot |st ) is the likelihood that observing image ot at state st . Put together, the posterior
estimation p(st |Ot ) can be computed efficiently. For object tracking, an ideal distribution
of p(st |Ot ) should peak at ot , i.e., st matching the observed object?s location ot . While
the integral in (1) predicts the regions where object is likely to appear given all the prior
observations, the observation model p(ot |st ) determines the most likely state that matches
the observation at time t.
In our formulation, p(ot |st ) measures the probability of observing ot as a sample being
generated by the target object class. Note that Ot is an image sequence and if the images
are acquired at high frame rate, it is expected that the difference between ot and ot?1
is small though object?s appearance might vary according to different of viewing angles,
illuminations, and possible self-deformation. Instead of adopting a complex static model
to learn p(ot |st ) for all possible ot , a simpler model suffices by adapting this model to
account for the appearance changes. In addition, since ot and ot?1 are most likely similar
and computing p(ot |st ) depends on p(ot?1 |st?1 ), the prior information p(ot?1 |st?1 ) can
be used to enhance the distinctiveness between the object and its background in p(ot |st ).
The idea of using an adaptive observation model for object tracking and then applying
discriminative analysis to better predict object location is the focus of the rest the paper. The
observation model we use is based on probabilistic principle component analysis (PPCA)
[10]. Object Tracking using PCA models have been well exploited in the computer vision
community [2]. Nevertheless, most existing tracking methods do not update the observation
models as time progresses. In this paper, we follow the work by Tipping and Bishop [10]
and propose an adaptive observation model based on PCA within a formal probabilistic
framework. Our result is a generalization of the conventional Fisher Linear Discriminant
with proper probabilistic interpretation.
3 A Discriminative Generative Observation Model
In this work, we track a target object based on its observations in the videos, i.e., ot . Since
the size of image region ot might change according to different st , we first convert ot to
a standard size and use it for tracking. In the following, we denote yt as the standardized
appearance vector of ot .
The dimensionality of the appearance vector yt is usually high. In our experiments, the
standard image size is a 19 ? 19 patch and thus yt is a 361-dimensional vector. We thus
model the appearance vector with a graphical model of low-dimensional latent variables.
3.1 A Generative Model with Latent Variables
A latent model relates a n-dimensional appearance vector y to a m-dimensional vector of
latent variables x:
y = Wx + ? +
(2)
where W is a n ? m projection matrix associating y and x, ? is the mean of y, and is
additive noise. As commonly assumed in factor analysis [1] and other graphical models [6],
the latent variables x are independent with unit variance, x ? N (0, Im ), where Im is the
m-dimensional identity matrix, and is zero mean Gaussian noise, ? N (0, ? 2 In ). Since
x and are both Gaussians, it follows that y is also a Gaussian distribution, y ? N (?, C),
where C = W W T + ? 2 I and In is an n-dimensional identity matrix. Together with (2),
we have a generative observation model:
p(ot |st ) = p(yt |W, ?, ) ? N (yt |?, W W T + ? 2 In )
(3)
This latent variable model follows the form of probabilistic principle component analysis,
and its parameters can be estimated from a set of examples [10]. Given a set of appearance
P
samples Y = {y1 , . . . , yN }, the covariance matrix of Y is denoted as S = N1 (y ?
?)(y ? ?)T . Let {?i |i = 1, . . . , N } be the eigenvalues of S arranged in descending order,
i.e., ?i ? ?j if i < j. Also, define the diagonal matrix ?m = diag(?1 , . . . , ?m ), and let
Um be the eigenvectors that corresponds to the eigenvalues in ?m . Tipping and Bishop
[10] show that the the maximum likelihood estimate of ?, W and can be obtained by
N
n
X
1 X
1
yi , W = Um (?m ? ? 2 Im )1/2 R, ? 2 =
?i
(4)
?=
N i=1
n ? m i=m+1
where R is an arbitrary m ? m orthogonal rotation matrix.
To model all possible appearance variations of a target object (due to pose, illumination
and view angle change), one could resort to a mixture of PPCA models. However, it not
only requires significant computation for estimating the model parameters but also leads
to other serious questions such as the number of components as well as under-fitting or
over-fitting. On the other hand, at any given time a linear PPCA model suffices to model
gradual appearance variation if the model is constantly updated. In this paper, we use a
single PPCA, and dynamically adapt the model parameters W , ?, and ? 2 to account for
appearance change.
3.1.1 Probability computation with Probabilistic PCA
Once the model parameters are known, we can compute the probability that a vector y is a
sample of this generative appearance model. From (4), the log-probability is computed by
1
(5)
L(W, ?, ? 2 ) = ? N log 2? + log |C| + y T C ?1 y
2
where y = y ? ?. Neglecting the constant terms, the log-probability is determined by
y T C ?1 y. Together with C = W W T + ? 2 In and (4), it follows that
1 T
T
T
yT C ?1 y = y T Um ??1
(6)
m Um y + 2 y (In ? Um Um )y
?
T
Here yT Um ??1
m Um y is the Mahalanobis distance of y in the subspace spanned by Um , and
T
T
y (In ? Um Um )y is the shortest distance from y to this subspace spanned by Um . Usually
? is set to a small value, and consequently the probability will be determined solely by the
distance to the subspace. However, the choice of ? is not trivial. From (6), if the ? is set to
a value much smaller than the actual one, the distance to the subspace will be favored and
ignore the contribution of Mahalanobis distance, thereby rendering an inaccurate estimate.
The choice of ? is even more critical in situations where the appearance changes dynamically and requires ? to be adjusted accordingly. This topic will be further examined in the
following section.
3.1.2 Online Learning of Probabilistic PCA
Unlike the analysis in the previous section where model parameters are estimated based on
a fixed set of training examples, our generative model has to learn and update its parameters
on line. Starting with a single example (the appearance of the tracked object in the first
video frame), our generative model constantly updates its parameters as new observations
arrive.
The equations for updating parameters are derived from (4). The update procedure of Um
and ?m is complicated since it involves the computations of eigenvalues and eigenvectors.
Here we use a forgetting factor ? to put more weights on the most recent data. Denote the
newly arrived samples at time t as Y = {y 1 , . . . , y M }, and assume the mean ? is fixed,
t
Um
and ?tm can be obtained by performing singular value decomposition (SVD) on
p
?
(7)
[ ?Um,t?1 (?m,t?1 )1/2 | (1 ? ?)Y ]
1/2
where Y = [y 1 ? ?, . . . , y M ? ?]. ?m,t and Um,t will contain the m-largest singular values
and the corresponding singular vectors respectively at time t. This update procedure can
be efficiently implemented using the R-SVD algorithm, e.g., [4] [7].
If the mean ? constantly changes, the above update procedure can not be applied. We
1/2
recently proposed a method [8] to compute SVD with correct updated mean in which ?m,t
and Um,t can be obtained by computing SVD on
p
p
h?
i
?Um,t?1 (?m,t?1 )1/2 (1 ? ?)Y (1 ? ?)?(?t?1 ? ?Y )
(8)
P
M
1
i
where Y = [y 1 ? ?Y , . . . , y M ? ?Y ] and ?Y = M
i=1 y . This formulation is similar to
the SVD computation with the fixed mean case, and the same incremental SVD algorithm
1/2
can be used to compute ?m,t and Um,t with an extra term shown in (8).
Computing and updating ? is more difficult than the form in (8). In the previous section,
we show that an inaccurate value of ? will severely affect probability estimates. In order
to have an accurate estimate of ? using (4), a large set of training examples is usually
required. Our generative model starts with a single example and gradually adapts the model
parameters. If we update ? based on (4), we will start with a very small value of ? since
there are only a few samples at our disposal at the outset, and the algorithm could quickly
lose track of the target because of an inaccurate probability estimate. Since the training
examples are not permanently stored in memory, ?i in (4) and consequently ? may not be
accurately updated if the number of drawn samples is insufficient. These constraints lead
us to develop a method that adaptively adjusts ? according to the newly arrived samples,
which will be explained in the next section.
3.2 Discriminative Generative Model
As is observed in Section 2, the object?s appearance at ot?1 and ot do not change much.
Therefore, we can use the observation at ot?1 to boost the likelihood measurement in ot .
That is, we draw a set samples (i.e., image patches) parameterized by {sit?1 |i = 1, ..., k}
in ot?1 that have large p(ot?1 |sit?1 ), but the low posterior p(sit?1 |Ot?1 ). These are treated
as the negative samples (i.e., samples that are not generated from the class of the target
object) that the generative model is likely to confuse at Ot .
Given a set samples Y 0 = {y 1 , . . . , y k } where y i is the appearance vector collected in ot?1
based on state parameter sit?1 , we want to find a linear projection V ? that projects Y 0 onto
a subspace such that the likelihood of Y 0 in the subspace is minimized. Let V be a p ? n
matrix and since p(y|W, ?, ?) is a Gaussian, p(V y|V, W, ?, ?) ? N (V ?, V CV T ) is a also
Gaussian. The log likelihood is computed by
k
(9)
L(V, W, ?, ?) = ? p log(2?) + log |V CV T | + tr((V CV T )?1 V S 0 V T )
2
P
where S 0 = k1 ki=1 (y i ? ?)(y i ? ?)T .
To facilitate the following analysis we first assume V projects Y 0 to a 1-D space, i.e., p = 1
and V = v T , and thus
k
vT S 0 v
T
L(V, W, ?, ?) = ?
log(2?) + log |v Cv| + T
(10)
2
v Cv
Note that v T Cv is the variance of the object samples in the projected space, and we need
to impose a constraint, e.g., v t Cv = 1, to ensure that the minimum likelihood solution of
v does not increase the variance in the projected space. Let v T Cv = 1, the optimization
problem becomes
vT S 0 v
v ? = arg max v T S 0 v = arg max T
(11)
v
v Cv
{v|v T Cv=1}
Thus, we obtain an equation exactly like the Fisher discriminant analysis for a binary classification problem. In (11), v is a projection that keeps the object?s samples in the projected
space close to the ? (with variance v T Cv = 1), while keeping negative samples in Y 0 away
from ?. The optimal value of v is the generalized eigenvector of S 0 and C that corresponds
to largest eigenvalue. In a general case, it follows that
|V S 0 V T |
(12)
V ? = arg max |V S 0 V T | = arg max
V
|V CV T |
{V CV T =I}
where V ? can be obtained by solving a generalized eigenvalue problem of S 0 and C. By
projecting observation samples onto a low dimensional subspace, we enhance the discriminative power of the generative model. In the meanwhile, we reduce the time required to
compute probabilities, which is also a critical improvement for real time applications like
object tracking.
3.2.1 Online Update of Discriminative Analysis
The computation of the projection matrix V depends on matrices C and S 0 . In section
3.1.2, we have shown the procedures to update C. The same procedures can be used to
P
P
update S 0 . Let ?Y 0 = k1 ki=1 y i and SY 0 = k1 ki=1 (y i ? ?Y 0 )(y i ? ?Y 0 )T ,
S0 =
k
1X i
(y ? ?)(y i ? ?)T = Sy + (? ? ?Y 0 )(? ? ?Y 0 )T
k i=1
(13)
Given S 0 and C, V is computed by solving a generalized eigenvalue problem. If we decompose S 0 = AT A and C = B T B, then we can find V more efficiently using generalized
singular value decomposition. Denote UY 0 and ?Y 0 as the SVD of SY 0 , it follows that by
1/2
1/2
letting A = [UY 0 ?Y 0 | (? ? ?Y 0 )]T and B = [Um ?m |? 2 I]T , we obtain S 0 = AT A and
T
C = B B.
As is detailed in [4] , V can be computed by first performing a QR factorization:
A
QA
=
R
(14)
B
QB
and computing the singular value decomposition of QA
(15)
QA = UA DA VAT
?1
, we then obtain V = R VA . The rank of A is usually small in vision applications, and V
can be computed efficiently, thereby facilitating tracking the process.
4 Proposed Tracking Algorithm
In this section, we summarize the proposed tracking algorithm and demonstrate how the
abovementioned learning and inference algorithms are incorporated for object tracking.
Our algorithm localizes the tracked object in each video frame using a rectangular window.
A state s is a length-5 vector, s = (x, y, ?, w, h), that parameterizes the windows position
(x, y), orientation (?) and width and height (w, h). The proposed algorithm is based on
maximum likelihood estimate (i.e., the most probable location of the object) given all the
observations up to that time instance, s?t = arg maxst p(st |Ot ).
We assume that state transition is a Gaussian distribution, i.e.,
p(st |st?1 ) ? N (st?1 , ?s )
(16)
where ?s is a diagonal matrix. According to this distribution, the tracker then draws N
samples St = {c1 , . . . , cN } which represent the possible locations of the target. Denote yti
as the appearance vector of ot , and Yt = {yt1 , . . . , ytN } as a set of appearance vectors that
corresponds to the set of state vectors St . The posterior probability that the tracked object
is at ci in video frame ot is then defined as
p(st = ci |Ot ) = ?p(yti |V, W, ?, ?)p(st = ci |s?t?1 )
(17)
where ? is a constant. Therefore, s?t = arg maxci ?St p(st = ci |Ot ).
Once s?t is determined, the corresponding observation yt? will be a new example to update
W and ?. Appearance vectors yti with large p(yti |V, W, ?, ?) but whose corresponding state
parameters ci are away from s?t will be used as new examples to update V .
Our tracking assumes o1 and s?1 are given (through object detection) and thus obtains the
first appearance vector y1 which in turn is used an the initial value of ?, but V and W are
unknown at the outset. When V and W are not available, our tracking algorithm is based
on template matching (with ? being the template). The matrix W is computed after a small
number of appearance vectors are observed. When W is available, we can then start to
compute and update V accordingly.
As mentioned in the Section 3.1.1, it is difficult to obtain an accurate estimate of ?. In our
tracking the system, we adaptively adjust ? according to ?m in W . We set ? be a fixed
fraction of the smallest eigenvalues in ?m . This will ensure the distance measurement
in (6) will not be biased to favor either the Mahalanobis distance in the subspace or the
distance to the subspace.
5 Experimental Results
We tested the proposed algorithm with numerous object tracking experiments. To examine whether our model is able to adapt and track objects in the dynamically changing environment, we recorded videos containing appearance deformation, large illumination change, and large pose variations. All the image sequences consist of 320 ? 240
pixel grayscale videos, recorded at 30 frames/second and 256 gray-levels per pixel. The
forgetting term is empirically selected as 0.85, and the batch size for update is set to 5
as a trade-off of computational efficiency as well as effectiveness of modeling appearance change due to fast motion. More experimental results and videos can be found at
http://www.ifp.uiuc.edu/?rlin1/adgm.html.
Figure 1: A target undergoes pose and lighting variation.
Figures 1 and 2 show snapshots of some tracking results enclosed with rectangular windows. There are two rows of images below each video frame. The first row shows the
sampled images in the current frame that have the largest likelihoods of being the target locations according our discriminative generative model. The second row shows the sample
images in the current video frame that are selected online for updating the discriminative
generative model.
The results in Figure 1 show the our method is able to track targets undergoing pose and
lighting change. Figure 2 shows tracking results where the object appearances change
significantly due to variation in pose and lighting as well as cast shadows. These experiments demonstrate that our tracking algorithm is able to follow objects even when there
is a large appearance change due to pose or lighting variation. We have also tested these
two sequences with conventional view-based eigentracker [2] or template-based method.
Empirical results show that such methods do not perform well as they do not update the
object representation to account for appearance change.
Figure 2: A target undergoes large lighting and pose variation with cast shadows.
6 Conclusion
We have presented a discriminative generative framework that generalizes the conventional
Fisher Linear Discriminant algorithm with a proper probabilistic interpretation. For object
tracking, we aim to find a discriminative generative model that best separates the target
class from the background. With a computationally efficient algorithm that constantly update this discriminative model as time progresses, our method adapts the discriminative
generative model to account for appearance variation of the target and background, thereby
facilitating the tracking task in different situations. Our experiments show that the proposed model is able to learn a discriminative generative model for tracking target objects
undergoing large pose and lighting changes. We also plan to apply the proposed method to
other problems that deal with non-stationary data stream in our future work.
References
[1] T. W. Anderson. An Introduction to Multivariate Statistical Analysis. Wiley, New York, 1984.
[2] M. J. Black and A. D. Jepson. Eigentracking: Robust matching and tracking of articulated objects using
view-based representation. In B. Buxton and R. Cipolla, editors, Proceedings of the Fourth European
Conference on Computer Vision, LNCS 1064, pp. 329?342. Springer Verlag, 1996.
[3] R. T. Collins and Y. Liu. On-line selection of discriminative tracking features. In Proceedings of the Ninth
IEEE International Conference on Computer Vision, volume 1, pp. 346?352, 2003.
[4] G. H. Golub and C. F. Van Loan. Matrix Computations. The Johns Hopkins University Press, 1996.
[5] M. Isard and A. Blake. Contour tracking by stochastic propagation of conditional density. In B. Buxton
and R. Cipolla, editors, Proceedings of the Fourth European Conference on Computer Vision, LNCS 1064,
pp. 343?356. Springer Verlag, 1996.
[6] M. I. Jordan, editor. Learning in Graphical Models. MIT Press, 1999.
[7] A. Levy and M. Lindenbaum. Sequential Karhunen-Loeve basis extraction and its application to images.
IEEE Transactions on Image Processing, 9(8):1371?1374, 2000.
[8] R.-S. Lin, D. Ross, J. Lim, and M.-H. Yang. Incremental subspace update with running mean.
Technical report, Beckman Institute, University of Illinois at Urbana-Champaign, 2004. available at
http://www.ifp.uiuc.edu/?rlin1/isuwrm.pdf.
[9] D. Ross, J. Lim, and M.-H. Yang. Adaptive probabilistic visual tracking with incremental subspace update.
In T. Pajdla and J. Matas, editors, Proceedings of the Eighth European Conference on Computer Vision,
LNCS 3022, pp. 470?482. Springer Verlag, 2004.
[10] M. E. Tipping and C. M. Bishop. Probabilistic principal component analysis. Journal of the Royal Statistical Society, Series B, 61(3):611?622, 1999.
| 2642 |@word gradual:1 covariance:1 decomposition:3 paid:1 thereby:4 tr:1 ytn:1 initial:1 liu:1 contains:2 series:1 existing:2 current:4 com:1 john:1 additive:1 eigentracking:1 wx:1 shape:1 update:19 stationary:1 generative:25 selected:2 isard:1 accordingly:2 beginning:1 toronto:2 honda:2 location:11 simpler:1 height:1 direct:1 fitting:2 acquired:1 forgetting:2 expected:2 indeed:1 examine:1 uiuc:4 ming:1 actual:1 window:3 ua:1 becomes:1 project:2 estimating:1 eigenvector:1 sung:1 exactly:1 um:20 classifier:2 unit:1 appear:1 yn:1 before:1 positive:4 dross:1 severely:2 solely:1 might:2 black:1 examined:1 dynamically:3 collect:2 factorization:1 uy:2 recursive:1 procedure:5 lncs:3 empirical:1 significantly:3 adapting:1 projection:5 matching:3 word:1 outset:2 lindenbaum:1 onto:2 close:1 selection:1 put:2 context:1 impossible:1 applying:1 descending:1 www:2 conventional:6 yt:9 attention:1 starting:1 rectangular:2 formulate:2 hsuan:1 assigns:2 adjusts:1 spanned:2 variation:12 updated:3 target:27 hypothesis:3 updating:3 predicts:2 observed:6 capture:1 region:4 trade:1 mentioned:1 environment:3 dynamic:2 trained:1 solving:2 efficiency:1 basis:1 articulated:1 fast:1 effective:1 monte:1 kp:1 detected:1 whose:2 favor:1 unseen:1 rlin1:3 online:3 sequence:3 eigenvalue:7 propose:1 flexibility:1 adapts:3 qr:1 produce:1 incremental:3 object:50 develop:3 pose:11 progress:5 implemented:1 c:1 predicted:1 involves:1 shadow:2 correct:1 filter:1 stochastic:1 viewing:3 premise:1 suffices:2 generalization:2 decompose:1 probable:1 im:3 adjusted:1 lying:1 tracker:9 blake:1 predict:3 vary:1 smallest:1 estimation:5 lose:1 beckman:1 ross:3 largest:3 hope:1 mit:1 gaussian:7 aim:3 encode:1 derived:1 focus:3 improvement:1 rank:1 indicates:1 likelihood:8 inference:2 inaccurate:3 unlikely:1 selects:1 pixel:2 unobservable:2 among:1 classification:2 orientation:2 denoted:1 favored:1 arg:6 html:1 plan:1 construct:1 once:3 extraction:1 sampling:1 future:1 minimized:1 report:1 serious:1 few:2 n1:1 detection:1 adjust:1 golub:1 mixture:1 accurate:2 ambient:1 beforehand:1 integral:1 neglecting:1 orthogonal:1 deformation:5 instance:1 modeling:1 markovian:1 cover:1 imperative:1 stored:1 varies:1 considerably:1 adaptively:3 st:35 density:3 referring:1 peak:1 international:1 probabilistic:14 off:1 enhance:2 together:3 quickly:1 hopkins:1 reflect:1 recorded:2 containing:1 resort:1 account:6 includes:1 depends:2 stream:1 view:3 observing:2 start:4 complicated:1 contribution:1 variance:5 largely:1 efficiently:4 sy:3 correspond:1 accurately:1 carlo:1 lighting:9 pp:4 static:4 sampled:1 ppca:4 newly:2 lim:3 knowledge:3 enumerates:1 improves:1 dimensionality:1 disposal:1 tipping:3 follow:2 daunting:1 formulation:3 done:2 though:2 arranged:1 anderson:1 stage:3 hand:1 propagation:1 undergoes:2 gray:1 facilitate:1 contain:1 assigned:1 deal:1 attractive:1 mahalanobis:3 self:2 width:1 generalized:4 pdf:1 arrived:2 demonstrate:2 motion:2 image:17 recently:1 ifp:2 rotation:1 empirically:1 tracked:4 volume:1 interpretation:5 significant:1 measurement:2 cv:13 illinois:2 moving:2 posterior:3 multivariate:1 recent:2 belongs:1 certain:1 verlag:3 binary:2 vt:2 yi:1 exploited:1 minimum:1 impose:1 shortest:1 relates:1 infer:2 reduces:1 champaign:1 technical:1 match:1 adapt:2 lin:2 myang:1 vat:1 va:1 prediction:1 vision:7 circumstance:1 represent:1 adopting:1 robotics:1 c1:1 background:9 addition:1 want:1 singular:5 ot:49 operate:1 rest:1 unlike:1 extra:1 biased:1 validating:1 effectiveness:1 jordan:1 yang:3 ideal:1 rendering:1 affect:1 associating:1 reduce:1 idea:1 tm:1 parameterizes:1 cn:1 whether:1 pca:4 render:1 york:1 detailed:1 eigenvectors:2 http:2 restricts:1 estimated:3 track:7 per:1 nevertheless:1 drawn:1 changing:3 utilize:1 fraction:1 convert:1 jongwoo:1 angle:4 parameterized:1 fourth:2 dst:1 arrive:1 patch:2 yt1:1 draw:2 ki:3 distinguish:1 constraint:2 ri:1 loeve:1 qb:1 performing:2 developing:1 according:6 belonging:2 smaller:1 explained:1 gradually:1 projecting:1 computationally:2 equation:3 previously:1 turn:1 mechanism:1 fail:2 letting:1 end:1 generalizes:2 gaussians:1 available:3 apply:2 away:2 alternative:1 robustness:1 batch:1 permanently:1 standardized:1 assumes:1 ensure:2 running:1 graphical:3 k1:3 society:1 matas:1 question:1 ruei:1 diagonal:2 abovementioned:2 subspace:12 distance:9 separate:3 topic:2 collected:1 discriminant:5 trivial:1 toward:1 assuming:1 kalman:1 o1:2 modeled:1 length:1 insufficient:1 difficult:2 pajdla:1 negative:8 reliably:1 proper:5 unknown:1 perform:1 observation:28 snapshot:1 urbana:1 situation:3 ever:2 incorporated:1 frame:9 y1:2 ninth:1 arbitrary:1 community:2 david:1 cast:2 required:2 adgm:1 learned:1 boost:1 qa:3 able:5 usually:4 perception:1 below:1 eighth:1 summarize:1 max:4 memory:1 video:9 royal:1 power:1 critical:2 treated:1 localizes:1 numerous:3 prior:5 determining:1 enclosed:1 buxton:2 s0:1 principle:2 editor:4 row:3 keeping:1 formal:1 institute:2 distinctiveness:1 template:3 van:1 transition:4 contour:1 commonly:1 adaptive:8 projected:3 transaction:1 obtains:1 ignore:1 keep:1 active:1 assumed:1 discriminative:23 grayscale:1 continuous:1 latent:6 decade:1 learn:6 robust:3 complex:1 meanwhile:1 european:3 domain:1 diag:1 da:1 jepson:1 noise:2 facilitating:3 wiley:1 position:2 levy:1 bishop:3 undergoing:3 sit:4 essential:1 consist:1 sequential:1 ci:5 illumination:5 confuse:1 karhunen:1 simply:2 appearance:32 likely:6 visual:2 tracking:40 cipolla:2 springer:3 corresponds:3 environmental:1 determines:2 constantly:5 conditional:1 goal:1 formulated:4 identity:2 consequently:3 fisher:5 yti:4 change:18 loan:1 specifically:1 determined:3 principal:2 experimental:3 svd:7 select:1 collins:1 tested:2 |
1,809 | 2,643 | Hierarchical Bayesian Inference in
Networks of Spiking Neurons
Rajesh P. N. Rao
Department of Computer Science and Engineering
University of Washington, Seattle, WA 98195
[email protected]
Abstract
There is growing evidence from psychophysical and neurophysiological
studies that the brain utilizes Bayesian principles for inference and decision making. An important open question is how Bayesian inference
for arbitrary graphical models can be implemented in networks of spiking neurons. In this paper, we show that recurrent networks of noisy
integrate-and-fire neurons can perform approximate Bayesian inference
for dynamic and hierarchical graphical models. The membrane potential
dynamics of neurons is used to implement belief propagation in the log
domain. The spiking probability of a neuron is shown to approximate the
posterior probability of the preferred state encoded by the neuron, given
past inputs. We illustrate the model using two examples: (1) a motion detection network in which the spiking probability of a direction-selective
neuron becomes proportional to the posterior probability of motion in
a preferred direction, and (2) a two-level hierarchical network that produces attentional effects similar to those observed in visual cortical areas
V2 and V4. The hierarchical model offers a new Bayesian interpretation
of attentional modulation in V2 and V4.
1
Introduction
A wide range of psychophysical results have recently been successfully explained using
Bayesian models [7, 8, 16, 19]. These models have been able to account for human responses in tasks ranging from 3D shape perception to visuomotor control. Simultaneously,
there is accumulating evidence from human and monkey experiments that Bayesian mechanisms are at work during visual decision making [2, 5]. The versatility of Bayesian models
stems from their ability to combine prior knowledge with sensory evidence in a rigorous
manner: Bayes rule prescribes how prior probabilities and stimulus likelihoods should be
combined, allowing the responses of subjects or neural responses to be interpreted in terms
of the resulting posterior distributions.
An important question that has only recently received attention is how networks of cortical neurons can implement algorithms for Bayesian inference. One powerful approach has
been to build on the known properties of population coding models that represent information using a set of neural tuning curves or kernel functions [1, 20]. Several proposals have
been made regarding how a probability distribution could be encoded using population
codes ([3, 18]; see [14] for an excellent review). However, the problem of implementing
general inference algorithms for arbitrary graphical models using population codes remains
unresolved (some encouraging initial results are reported in Zemel et al., this volume). An
alternate approach advocates performing Bayesian inference in the log domain such that
multiplication of probabilities is turned into addition and division to subtraction, the latter
operations being easier to implement in standard neuron models [2, 5, 15] (see also the
papers by Deneve and by Yu and Dayan in this volume). For example, a neural implementation of approximate Bayesian inference for a hidden Markov model was investigated in
[15]. The question of how such an approach could be generalized to spiking neurons and
arbitrary graphical models remained open.
In this paper, we propose a method for implementing Bayesian belief propagation in networks of spiking neurons. We show that recurrent networks of noisy integrate-and-fire
neurons can perform approximate Bayesian inference for dynamic and hierarchical graphical models. In the model, the dynamics of the membrane potential is used to implement
on-line belief propagation in the log domain [15]. A neuron?s spiking probability is shown
to approximate the posterior probability of the preferred state encoded by the neuron, given
past inputs. We first show that for a visual motion detection task, the spiking probability
of a direction-selective neuron becomes proportional to the posterior probability of motion
in the neuron?s preferred direction. We then show that in a two-level network, hierarchical
Bayesian inference [9] produces responses that mimic the attentional effects seen in visual
cortical areas V2 and V4.
2
Modeling Networks of Noisy Integrate-and-Fire Neurons
2.1 Integrate-and-Fire Model of Spiking Neurons
We begin with a recurrently-connected network of integrate-and-fire (IF) neurons receiving
feedforward inputs denoted by the vector I. The membrane potential of neuron i changes
according to:
X
X
dvi
?
= ?vi +
wij Ij +
uij vj0
(1)
dt
j
j
where ? is the membrane time constant, Ij denotes the synaptic current due to input neuron
j, wij represents the strength of the synapse from input j to recurrent neuron i, vj0 denotes the synaptic current due to recurrent neuron j, and uij represents the corresponding
synaptic strength. If vi crosses a threshold T , the neuron fires a spike and vi is reset to the
potential vreset . Equation 1 can be rewritten in discrete form as:
X
X
vi (t + 1) = vi (t) + (?vi (t) +
wij Ij (t)) +
uij vj0 (t))
(2)
j
i.e.
vi (t + 1)
=
X
wij Ij (t) +
X
j
u0ij vj0 (t)
(3)
j
j
where is the integration rate, u0ii = 1 + (uii ? 1) and for i 6= j, u0ij = uij .
A more general integrate-and-fire model that takes into account some of the effects of nonlinear filtering in dendrites can be obtained by generalizing Equation 3 as follows:
X
X
vi (t + 1) = f
wij Ij (t) + g
u0ij vj0 (t)
(4)
j
j
where f and g model potentially different dendritic filtering functions for feedforward and
recurrent inputs.
2.2 Stochastic Spiking in Noisy IF Neurons
To model the effects of background inputs and the random openings of membrane channels,
one can add a Gaussian white noise term to the right hand side of Equations 3 and 4. This
makes the spiking of neurons in the recurrent network stochastic. Plesser and Gerstner [13]
and Gerstner [4] have shown that under reasonable assumptions, the probability of spiking
in such noisy neurons can be approximated by an ?escape function? (or hazard function)
that depends only on the distance between the (noise-free) membrane potential v i and the
threshold T . Several different escape functions were studied. Of particular interest to the
present paper is the following exponential function for spiking probability suggested in [4]
for noisy integrate-and-fire networks:
P (neuron i spikes at time t) = ke(vi (t)?T )/c
(5)
where k and c are arbitrary constants. We used a model that combines Equations 4 and 5
to generate spikes, with an absolute refractory period of 1 time step.
3
Bayesian Inference using Spiking Neurons
3.1 Inference in a Single-Level Model
We first consider on-line belief propagation in a single-level dynamic graphical model and
show how it can be implemented in spiking networks. The graphical model is shown in
Figure 1A and corresponds to a classical hidden Markov model. Let ?(t) represent the
hidden state of a Markov model at time t with transition probabilities given by P (?(t) =
?i |?(t ? 1) = ?j ) = P (?it |?jt?1 ) for i, j = 1 . . . N . Let I(t) be the observable output
governed by the probabilities P (I(t)|?(t)). Then, the forward component of the belief
propagation algorithm [12] prescribes the following ?message? for state i from time step t
to t + 1:
X
mt,t+1
= P (I(t)|?it )
P (?it |?jt?1 )mt?1,t
(6)
i
j
j
m0,1
i
If
= P (?i ) (the prior distribution over states), then it is easy to show using Bayes
rule that mt,t+1
= P (?it , I(t), . . . , I(1)). If the probabilities are normalized at each update
i
step:
X
P (?it |?jt?1 )mt?1,t
/nt?1,t
(7)
mt,t+1
= P (I(t)|?it )
j
i
j
P
where nt?1,t = j mt?1,t
, then the message becomes equal to the posterior probability
j
of the state and current input, given all past inputs:
mt,t+1
= P (?it , I(t)|I(t ? 1), . . . , I(1))
i
(8)
3.2 Neural Implementation of the Inference Algorithm
By comparing the membrane potential equation (Eq. 4) with the on-line belief propagation equation (Eq. 7), it is clear that the first equation can implement the second if belief
propagation is performed in the log domain [15], i.e., if:
vi (t + 1) ? log mt,t+1
i
X
f
wij Ij (t) = log P (I(t)|?it )
(9)
(10)
j
g
X
j
u0ij vj0 (t)
=
log(
X
P (?it |?jt?1 )mt?1,t
/nt?1,t )
j
(11)
j
In this model, the dendritic filtering functions f and g approximate the logarithm function1 , the synaptic currents Ij (t) and vj0 (t) are approximated by the corresponding instantaneous firing rates, and the recurrent synaptic weights u0ij encode the transition probabilities P (?it |?jt?1 ). Normalization by nt?1,t is implemented by subtracting log nt?1,t using
inhibition.
1
An alternative approach, which was also found to yield satisfactory results, is to approximate the
log-sum with a linear weighted sum [15], the weights being chosen to minimize the approximation
error.
t
t+1
?t
t+1
? t+1
? t+1
1
?t
1
I(t+1)
I(t)
? t+1
2
?t
2
t
I(t)
B
A
I(t)
I(t+1)
I(t)
D
C
Figure 1: Graphical Models and their Neural Implementation. (A) Single-level dynamic graph-
ical model. Each circle represents a node denoting the state variable ? t which can take on values
?1 , . . . , ?N . (B) Recurrent network for implementing on-line belief propagation for the graphical
model in (A). Each circle represents a neuron encoding a state ? i . Arrows represent synaptic connections. The probability distribution over state values at each time step is represented by the entire
population. (C) Two-level dynamic graphical model. (D) Two-level network for implementing online belief propagation for the graphical model in (C). Arrows represent synaptic connections in the
direction pointed by the arrow heads. Lines without arrow heads represent bidirectional connections.
Finally, since the membrane potential vi (t + 1) is assumed to be proportional to log mt,t+1
i
(Equation 9), we have:
vi (t + 1) = c log mt,t+1
+T
(12)
i
for some constants c and T . For noisy integrate-and-fire neurons, we can use Equation 5 to
calculate the probability of spiking for each neuron i as:
P (neuron i spikes at time t + 1)
? e(vi (t+1)?T )/c
= e
log mt,t+1
i
=
mt,t+1
i
(13)
(14)
Thus, the probability of spiking (or equivalently, the instantaneous firing rate) for neuron i
in the recurrent network is directly proportional to the posterior probability of the neuron?s
preferred state and the current input, given all past inputs. Figure 1B illustrates the singlelevel recurrent network model that implements the on-line belief propagation equation 7.
3.3 Hierarchical Inference
The model described above can be extended to perform on-line belief propagation and
inference for arbitrary graphical models. As an example, we describe the implementation
for the two-level hierarchical graphical model in Figure 1C.
As in the case of the 1-level dynamic model, we define the following ?messages? within a
(message from state i to other states at level 1
particular level and between levels: mt,t+1
1,i
from time step t to t + 1), mt1?2,i (?feedforward? message from state i at level 1 sent to
(message from state i to other states at level 2 from time step t
level 2 at time t), mt,t+1
2,i
to t + 1), and mt2?1,i (?feedback? message from state i at level 2 sent to level 1 at time
t). Each of these messages can be calculated based on an on-line version of loopy belief
propagation [11] for the multiply connected two-level graphical model in Figure 1C:
XX
t?1
t
t
t
mt1?2,i =
P (?1,k
|?2,i
, ?1,j
)mt?1,t
(15)
1,j P (I(t)|?1,k )
j
mt2?1,i
=
X
j
k
t?1
t
P (?2,i
|?2,j
)mt?1,t
2,j
(16)
mt,t+1
1,i
t
)
= P (I(t)|?1,i
XX
j
mt,t+1
2,i
X
= mt1?2,i
t?1
t
t
|?2,j
, ?1,k
P (?1,i
)mt2?1,j mt?1,t
1,k
(17)
k
t?1
t
P (?2,i
|?2,j
)mt?1,t
2,j
(18)
j
Note the similarity between the last equation and the equation for the single-level model
(Equation 6). The equations above can be implemented in a 2-level hierarchical recurrent
network of integrate-and-fire neurons in a manner similar to the 1-level case. We assume
that neuron i in level 1 encodes ?1,i as its preferred state while neuron i in level 2 encodes ?2,i . We also assume specific feedforward and feedback neurons for computing and
conveying mt1?2,i and mt2?1,i respectively.
Taking the logarithm of both sides of Equations 17 and 18, we obtain equations that can
be computed using the membrane potential dynamics of integrate-and-fire neurons (Equation 4). Figure 1D illustrates the corresponding two-level hierarchical network. A modification needed to accommodate Equation 17 is to allow bilinear interactions between synaptic
inputs, which changes Equation 4 to:
X
XX
wij Ij (t) + g
(19)
vi (t + 1) = f
u0ijk vj0 (t)xk (t)
j
j
k
Multiplicative interactions between synaptic inputs have previously been suggested by several authors (e.g., [10]) and potential implementations based on active dendritic interactions
have been explored. The model suggested here utilizes these multiplicative interactions
within dendritic branches, in addition to a possible logarithmic transform of the signal before it sums with other signals at the soma. Such a model is comparable to recent models
of dendritic computation (see [6] for more details).
4
Results
4.1 Single-Level Network: Probabilistic Motion Detection and Direction Selectivity
We first tested the model in a 1D visual motion detection task [15]. A single-level recurrent
network of 30 neurons was used (see Figure 1B). Figure 2A shows the feedforward weights
for neurons 1, . . . , 15: these were recurrently connected to encode transition probabilities biased for rightward motion as shown in Figure 2B. Feedforward weights for neurons
16, . . . , 30 were identical to Figure 2A but their recurrent connections encoded transition
probabilities for leftward motion (see Figure 2B). As seen in Figure 2C, neurons in the
network exhibited direction selectivity. Furthermore, the spiking probability of neurons
reflected the posterior probabilities over time of motion direction at a given location (Figure 2D), suggesting a probabilistic interpretation of direction selective spiking responses in
visual cortical areas such as V1 and MT.
4.2
Two-Level Network: Spatial Attention as Hierarchical Bayesian Inference
We tested the two-level network implementation (Figure 1D) of hierarchical Bayesian inference using a simple attention task previously used in primate studies [17]. In an input
image, a vertical or horizontal bar could occur either on the left side, right side, or both
sides (see Figure 3). The corresponding 2-level generative model consisted of two states
at level 2 (left or right side) and four states at level 1: vertical left, horizontal left, vertical
right, horizontal right. Each of these states was encoded by a neuron at the respective level.
The feedforward connections at level 1 were chosen to be vertically or horizontally oriented
Gabor filters localized to the left or right side of the image. Since the experiment used static
images, the recurrent connections at each level implemented transition probabilities close
to 1 for the same state and small random values for other states. The transition probabilities
t?1
t
t
P (?1,k
|?2,i
, ?1,j
) were chosen such that for ?2t = left side, the transition probabilities for
?t
w15
w1
1
0.30.3
15
30
1
0.2
0.2
?
t+1
15
0.1
0.1
00
0
1
5
10
15
20
20
10
Spatial Location (pixels)
25
30
30
Rightward
30
Leftward
A
Rightward Motion
B
Leftward Motion
Neuron
Rightward Motion
Leftward Motion
0.5
0.5
8
0.4
0.4
0.3
0.3
0.2
0.2
0.1
0.1
0
10
15
20
25
0
0.5
0.4
0.3
0.2
0.2
0.1
0.1
5
10
15
20
25
0
0.5
0.5
0.4
0.4
0.3
0.3
0.2
0.2
0.1
0.1
0
C
10
0.3
0
12
5
0.5
0.4
5
10
15
20
25
0
5
10
15
20
25
30
5
10
15
20
25
30
5
10
15
20
25
30
D
Figure 2: Responses from the Single-Level Motion Detection Network. (A) Feedforward weights
for neurons 1, . . . , 15 (rightward motion selective neurons). Feedforward weights for neurons
16, . . . , 30 (leftward motion selective) are identical. (B) Recurrent weights encoding the transition
probabilities P (?it+1 |?jt ) for i, j = 1, . . . , 30. Probability values are proportional to pixel brightness.
(C) Spiking responses of three of the first 15 neurons in the recurrent network (neurons 8, 10, and
12). As is evident, these neurons have become selective for rightward motion as a consequence of
the recurrent connections specified in (B). (D) Posterior probabilities over time of motion direction
(at a given location) encoded by the three neurons for rightward and leftward motion.
states ?1t coding for the right side were set to values close to zero (and vice versa, for ? 2t =
right side). As shown in Figure 3, the response of a neuron at level 1 that, for example,
prefers a vertical edge on the right mimics the response of a V4 neuron with and without
attention (see figure caption for more details). The initial setting of the priors at level 2 is
the crucial determinant of attentional modulation in level 1 neurons, suggesting that feedback from higher cortical areas may convey task-specific priors that are integrated into V4
responses.
5
Discussion and Conclusions
We have shown that recurrent networks of noisy integrate-and-fire neurons can perform
approximate Bayesian inference for single- and multi-level dynamic graphical models. The
model suggests a new interpretation of the spiking probability of a neuron in terms of the
posterior probability of the preferred state encoded by the neuron, given past inputs. We
illustrated the model using two problems: inference of motion direction in a single-level
network and hierarchical inference of object identity at an attended visual location in a twolevel network. In the first case, neurons generated direction-selective spikes encoding the
probability of motion in a particular direction. In the second case, attentional effects similar
to those observed in primate cortical areas V2 and V4 emerged as a result of imposing
appropriate priors at the highest level.
The results obtained thus far are encouraging but several important questions remain. How
does the approach scale to more realistic graphical models? The two-level model explored
in this paper assumed stationary objects, resulting in simplified dynamics for the two levels
in our recurrent network. Experiments are currently underway to test the robustness of the
proposed model when richer classes of dynamics are introduced at the different levels. An-
A
B
30
Spikes/second
Spikes/second
30
Ref Att Away
25
20
15
0
5
10
15
20
20
15
10
0
Time steps from stim onset
C
Pair Att Away
25
5
10
15
20
Time steps from stim onset
D
Spikes/second
30
Pair Att Ref
25
20
15
0
5
10
15
20
Time steps from stim onset
Figure 3: Responses from the Two-Level Hierarchical Network. (A) Top panel: Input image
(lasting the first 15 time steps) containing a vertical bar (?Reference?) on the right side. Each input was convolved with a retinal spatiotemporal filter. Middle: Three sample spike trains from the
1st level neuron whose preferred stimulus was a vertical bar on the right side. Bottom: Posterior
probability of a vertical bar (= spiking probability or instantaneous firing rate of the neuron) plotted
over time. (B) Top panel: An input containing two stimuli (?Pair?). Below: Sample spike trains and
posterior probability for the same neuron as in (A). (C) When ?attention? is focused on the right side
(depicted by the white oval) by initializing the prior probability encoded by the 2nd level right-coding
neuron at a higher value than the left-coding neuron, the firing rate for the 1st level neuron in (A) increases to a level comparable to that in (A). (D) Responses from a neuron in primate area V4 without
attention (top panel, Ref Att Away and Pair Att Away; compare with (A) and (B)) and with attention
(bottom panel, Pair Att Ref; compare with (C)) (from [17]). Similar responses are seen in V2 [17].
other open question is how active dendritic processes could support probabilistic integration
of messages from local, lower-level, and higher-level neurons, as suggested in Section 3.
We intend to investigate this question using biophysical (compartmental) models of cortical
neurons. Finally, how can the feedforward, feedback, and recurrent synaptic weights in the
networks be learned directly from input data? We hope to investigate this question using
biologically-plausible approximations to the expectation-maximization (EM) algorithm.
Acknowledgments. This research was supported by grants from ONR, NSF, and the Packard Foundation. I am grateful to Wolfram Gerstner, Michael Shadlen, Aaron Shon, Eero Simoncelli, and Yair
Weiss for discussions on topics related to this paper.
References
[1] C. H. Anderson and D. C. Van Essen. Neurobiological computational systems. In Computational Intelligence: Imitating Life, pages 213?222. New York, NY: IEEE Press, 1994.
[2] R. H. S. Carpenter and M. L. L. Williams. Neural computation of log likelihood in control of
saccadic eye movements. Nature, 377:59?62, 1995.
[3] S. Deneve and A. Pouget. Bayesian estimation by interconnected neural networks (abstract no.
237.11). Society for Neuroscience Abstracts, 27, 2001.
[4] W. Gerstner. Population dynamics of spiking neurons: Fast transients, asynchronous states, and
locking. Neural Computation, 12(1):43?89, 2000.
[5] J. I. Gold and M. N. Shadlen. Neural computations that underlie decisions about sensory stimuli.
Trends in Cognitive Sciences, 5(1):10?16, 2001.
[6] M. Hausser and B. Mel. Dendrites: bug or feature? Current Opinion in Neurobiology, 13:372?
383, 2003.
[7] D. C. Knill and W. Richards. Perception as Bayesian Inference. Cambridge, UK: Cambridge
University Press, 1996.
[8] K. P. K?ording and D. Wolpert. Bayesian integration in sensorimotor learning. Nature, 427:244?
247, 2004.
[9] T. S. Lee and D. Mumford. Hierarchical Bayesian inference in the visual cortex. Journal of the
Optical Society of America A, 20(7):1434?1448, 2003.
[10] B. W. Mel. NMDA-based pattern discrimination in a modeled cortical neuron. Neural Computation, 4(4):502?517, 1992.
[11] K. Murphy, Y. Weiss, and M. Jordan. Loopy belief propagation for approximate inference: An
empirical study. In Proceedings of UAI (Uncertainty in AI), pages 467?475. 1999.
[12] J. Pearl. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. Morgan Kaufmann, San Mateo, CA, 1988.
[13] H. E. Plesser and W. Gerstner. Noise in integrate-and-fire neurons: From stochastic input to
escape rates. Neural Computation, 12(2):367?384, 2000.
[14] A. Pouget, P. Dayan, and R. S. Zemel. Inference and computation with population codes.
Annual Review of Neuroscience, 26:381?410, 2003.
[15] R. P. N. Rao. Bayesian computation in recurrent neural circuits. Neural Computation, 16(1):1?
38, 2004.
[16] R. P. N. Rao, B. A. Olshausen, and M. S. Lewicki. Probabilistic Models of the Brain: Perception
and Neural Function. Cambridge, MA: MIT Press, 2002.
[17] J. H. Reynolds, L. Chelazzi, and R. Desimone. Competitive mechanisms subserve attention in
macaque areas V2 and V4. Journal of Neuroscience, 19:1736?1753, 1999.
[18] M. Sahani and P. Dayan. Doubly distributional population codes: Simultaneous representation
of uncertainty and multiplicity. Neural Computation, 15:2255?2279, 2003.
[19] Y. Weiss, E. P. Simoncelli, and E. H. Adelson. Motion illusions as optimal percepts. Nature
Neuroscience, 5(6):598?604, 2002.
[20] R. S. Zemel, P. Dayan, and A. Pouget. Probabilistic interpretation of population codes. Neural
Computation, 10(2):403?430, 1998.
| 2643 |@word determinant:1 middle:1 version:1 nd:1 open:3 brightness:1 attended:1 accommodate:1 initial:2 att:6 denoting:1 ording:1 reynolds:1 past:5 current:6 comparing:1 nt:5 realistic:1 shape:1 update:1 discrimination:1 stationary:1 generative:1 intelligence:1 xk:1 wolfram:1 node:1 location:4 become:1 doubly:1 combine:2 advocate:1 manner:2 growing:1 multi:1 brain:2 encouraging:2 becomes:3 begin:1 xx:3 panel:4 circuit:1 interpreted:1 monkey:1 uk:1 control:2 grant:1 underlie:1 before:1 engineering:1 vertically:1 local:1 consequence:1 bilinear:1 encoding:3 firing:4 modulation:2 studied:1 mateo:1 suggests:1 range:1 acknowledgment:1 implement:6 illusion:1 area:7 empirical:1 gabor:1 close:2 accumulating:1 williams:1 attention:8 focused:1 ke:1 pouget:3 rule:2 population:8 caption:1 trend:1 approximated:2 richards:1 distributional:1 observed:2 bottom:2 initializing:1 calculate:1 connected:3 movement:1 highest:1 locking:1 dynamic:13 prescribes:2 grateful:1 division:1 rightward:7 represented:1 america:1 train:2 fast:1 describe:1 zemel:3 visuomotor:1 whose:1 encoded:8 emerged:1 richer:1 plausible:2 compartmental:1 ability:1 transform:1 noisy:8 online:1 biophysical:1 propose:1 subtracting:1 interaction:4 interconnected:1 unresolved:1 reset:1 turned:1 gold:1 bug:1 seattle:1 produce:2 object:2 illustrate:1 recurrent:21 ij:8 received:1 eq:2 implemented:5 c:1 direction:13 filter:2 stochastic:3 human:2 transient:1 opinion:1 implementing:4 dendritic:6 m0:1 estimation:1 currently:1 vice:1 successfully:1 weighted:1 hope:1 mit:1 gaussian:1 encode:2 likelihood:2 rigorous:1 am:1 inference:25 dayan:4 plesser:2 entire:1 integrated:1 hidden:3 uij:4 ical:1 selective:7 wij:7 pixel:2 denoted:1 spatial:2 integration:3 equal:1 washington:2 identical:2 represents:4 yu:1 adelson:1 mimic:2 stimulus:4 intelligent:1 escape:3 opening:1 oriented:1 simultaneously:1 u0ii:1 murphy:1 fire:13 versatility:1 detection:5 interest:1 message:9 investigate:2 essen:1 multiply:1 rajesh:1 edge:1 desimone:1 respective:1 logarithm:2 circle:2 plotted:1 modeling:1 rao:4 maximization:1 loopy:2 reported:1 spatiotemporal:1 combined:1 st:2 v4:8 probabilistic:6 receiving:1 lee:1 michael:1 w1:1 containing:2 cognitive:1 account:2 potential:9 suggesting:2 retinal:1 coding:4 vi:14 depends:1 onset:3 performed:1 multiplicative:2 competitive:1 bayes:2 minimize:1 kaufmann:1 percept:1 yield:1 conveying:1 bayesian:23 simultaneous:1 synaptic:10 sensorimotor:1 static:1 knowledge:1 nmda:1 bidirectional:1 higher:3 dt:1 reflected:1 response:13 wei:3 synapse:1 anderson:1 furthermore:1 hand:1 horizontal:3 nonlinear:1 propagation:13 vj0:8 olshausen:1 effect:5 normalized:1 consisted:1 satisfactory:1 illustrated:1 white:2 during:1 mel:2 generalized:1 evident:1 motion:22 reasoning:1 ranging:1 image:4 instantaneous:3 recently:2 spiking:23 mt:22 twolevel:1 refractory:1 function1:1 volume:2 interpretation:4 versa:1 imposing:1 cambridge:3 ai:1 tuning:1 subserve:1 pointed:1 similarity:1 cortex:1 inhibition:1 add:1 posterior:12 recent:1 leftward:6 selectivity:2 onr:1 life:1 seen:3 morgan:1 subtraction:1 period:1 signal:2 branch:1 simoncelli:2 stem:1 offer:1 cross:1 hazard:1 expectation:1 represent:5 kernel:1 normalization:1 proposal:1 background:1 addition:2 crucial:1 biased:1 exhibited:1 subject:1 sent:2 jordan:1 feedforward:10 easy:1 regarding:1 york:1 prefers:1 clear:1 generate:1 nsf:1 neuroscience:4 discrete:1 four:1 soma:1 threshold:2 v1:1 mt2:4 deneve:2 graph:1 sum:3 powerful:1 uncertainty:2 reasonable:1 utilizes:2 decision:3 comparable:2 chelazzi:1 annual:1 strength:2 occur:1 uii:1 encodes:2 performing:1 optical:1 department:1 according:1 alternate:1 membrane:9 remain:1 em:1 making:2 modification:1 primate:3 lasting:1 biologically:1 explained:1 multiplicity:1 imitating:1 equation:19 remains:1 previously:2 mechanism:2 needed:1 operation:1 rewritten:1 hierarchical:15 v2:6 appropriate:1 away:4 alternative:1 robustness:1 yair:1 convolved:1 denotes:2 top:3 graphical:16 build:1 classical:1 society:2 psychophysical:2 intend:1 question:7 spike:10 mumford:1 saccadic:1 distance:1 attentional:5 topic:1 stim:3 code:5 modeled:1 u0ij:5 equivalently:1 potentially:1 implementation:6 perform:4 allowing:1 vertical:7 neuron:73 markov:3 extended:1 neurobiology:1 head:2 arbitrary:5 introduced:1 pair:5 specified:1 connection:7 hausser:1 learned:1 pearl:1 macaque:1 able:1 suggested:4 bar:4 below:1 perception:3 pattern:1 packard:1 belief:13 eye:1 mt1:4 vreset:1 sahani:1 prior:7 review:2 multiplication:1 underway:1 proportional:5 filtering:3 localized:1 foundation:1 integrate:12 shadlen:2 principle:1 dvi:1 supported:1 last:1 free:1 asynchronous:1 side:13 allow:1 wide:1 taking:1 absolute:1 van:1 curve:1 feedback:4 cortical:8 transition:8 calculated:1 sensory:2 forward:1 made:1 author:1 san:1 simplified:1 far:1 approximate:9 observable:1 preferred:8 neurobiological:1 active:2 uai:1 assumed:2 eero:1 channel:1 nature:3 ca:1 dendrite:2 excellent:1 investigated:1 gerstner:5 domain:4 arrow:4 noise:3 w15:1 knill:1 ref:4 convey:1 carpenter:1 ny:1 exponential:1 governed:1 remained:1 specific:2 jt:6 recurrently:2 explored:2 evidence:3 illustrates:2 easier:1 wolpert:1 generalizing:1 logarithmic:1 depicted:1 neurophysiological:1 visual:8 horizontally:1 lewicki:1 shon:1 corresponds:1 ma:1 identity:1 change:2 oval:1 aaron:1 support:1 latter:1 tested:2 |
1,810 | 2,644 | A Temporal Kernel-Based Model for Tracking
Hand-Movements from Neural Activities
Lavi Shpigelman12 Koby Crammer1 Rony Paz23 Eilon Vaadia23 Yoram Singer1
1
School of computer Science and Engineering
2
Interdisciplinary Center for Neural Computation
3
Dept. of Physiology, Hadassah Medical School
The Hebrew University Jerusalem, 91904, Israel
Email for correspondance: [email protected]
Abstract
We devise and experiment with a dynamical kernel-based system for
tracking hand movements from neural activity. The state of the system
corresponds to the hand location, velocity, and acceleration, while the
system?s input are the instantaneous spike rates. The system?s state dynamics is defined as a combination of a linear mapping from the previous
estimated state and a kernel-based mapping tailored for modeling neural
activities. In contrast to generative models, the activity-to-state mapping
is learned using discriminative methods by minimizing a noise-robust
loss function. We use this approach to predict hand trajectories on the
basis of neural activity in motor cortex of behaving monkeys and find
that the proposed approach is more accurate than both a static approach
based on support vector regression and the Kalman filter.
1
Introduction
The paper focuses on the problem of tracking hand movements, which constitute smooth
spatial trajectories, from spike trains of a neural population. We do so by devising a dynamical system which employs a tailored kernel for spike trains along with a linear mapping
corresponding to the states? dynamics. Consider a situation where a subject performs free
hand movements during a task that requires accurate space and time precision. In the lab,
it may be a constrained reaching task while in real life it may be an every day task such as
eating. We wish to track the hand position given only spike trains from a recorded neural
population. The rationale of such an undertaking is two fold. First, this task can be viewed
as a stem towards the development of a Brain Machine Interface (BMI) which gradually and
rapidly become a possible future solution for the motor disabled patients. Recent studies of
BMIs [13, 3, 10] (being on-line and feedback enabled) show that a relatively small number
of cortical units can be used to move a cursor or a robot effectively, even without generation of hand movements and that training of the subjects improves the overall success of
the BMIs. Second, an open loop (off-line) movement decoding (see e.g. [7, 1, 15, 11, 8]),
while inappropriate for BMIs, is computationally less expensive, easier to implement and
allows repeated analysis thus providing a handle to understandings of neural computations
in the brain.
Early studies [6] showed that the direction of arm movement is reflected by the population
vector of preferred directions weighted by current firing rates, suggesting that intended
movement is encoded in the firing rate which, in turn, is modulated by the angle between a
unit?s preferred direction (PD) and the intended direction. This linear regression approach
is still prevalent and is applied, with some variation of the learning methods, in closed and
open loop settings. There is relatively little work on the development of dedicated nonlinear
methods.
Both movement and neural activity are dynamic and can therefore be naturally modeled by
dynamical systems. Filtering methods often employ generative probabilistic models such
as the well known Kalman filter [16] or more neurally specialized models [1] in which a
cortical unit?s spike count is generated by a probability function of its underlying firing
rate which is tuned to movement parameters. The movement, being a smooth trajectory,
is modeled as a linear transition with (typically additive Gaussian) noise. These methods
have the advantage of being aware of the smooth nature of movement and provide models
of what neurons are tuned to. However, the requirement of describing a neural population?s
firing probability as a function of movement state is hard to satisfy without making costly
assumptions. The most prominent is the assumption of statistical independence of cells
given the movement.
Kernel based methods have been shown to achieve state of the art results in many application domains. Discriminative kernel methods, such as Support Vector Regression (SVR)
forgo the task of modeling neuronal tuning functions. Furthermore, the construction of
kernel induced feature spaces, lends itself to efficient implementation of distance measures
over spike trains that are better suited to comparing two neural population trajectories than
the Euclidean distance in the original space of spike counts per bins [11, 5]. However,
SVR is a ?static? method that does not take into account the smooth dynamics of the predicted movement trajectory which imposes a statistical dependency between consecutive
examples.
This paper introduces a kernel based regression method that incorporates linear dynamics
of the predicted trajectories. In Sec. 2 we formally describe the problem setting. We introduce the movement tracking model and the associated learning framework in Sec. 3. The
resulting learning problem yields a new kernel for linear dynamical systems. We provide
an efficient calculation of this kernel and describe our dual space optimization method for
solving the learning problem. The experimental method is presented in Sec. 4. Results,
underscoring the merits of our algorithm are provided in Sec. 5 and conclusions are given
in Sec. 6.
2
Problem Setting
Our training set contains m trials. Each trial (typically indexed by i or j) consists of a pair
tiend
of movement and neural recordings, designated by Yi , Oi . Yi = yti t=1
is a time
d
i
series of movement state values and yt ? R is the movement state vector at time t in
trial i. We are interested in reconstructing position, however, for better modeling, yti may
be a vector of position, velocity and acceleration (as is the case in Sec. 4). This trajectory is
tiend
observed during model learning and is the inference target. Oi = {ot }t=1
is a time series
of neural spike counts and oit ? Rq is a vector
of spike counts from q cortical units at time
t. We wish to learn a function zit = f Oi1:t that is a good estimate (in a sense formalized
in the sequel) of the movement yti . Thus, f is a causal filtering method.
We confine ourselves to a causal setting since we plan to apply the proposed method in a
closed loop scenario where real-time output is required. The partition into separate trajectories is a natural one in a setting where a session is divided into many trials, each consisting
of one attempt at accomplishing the basic task (such as reaching movements to displayed
targets). In tasks that involve no hitting of objects, hand movements are typically smooth.
End point movement in small time steps is loosely approximated as having constant acceleration. On the other hand, neural spike counts (which are typically measured in bins
of 50 ? 100ms) vary greatly from one time step to the next. In summary, our goal is to
devise a dynamic mapping from sequences of neural activities ending at a given time to the
instantaneous hand movement characterization (location, velocity, and acceleration).
3
Movement Tracking Algorithm
Our regression method is defined as follows: given a series O ? Rq?tend of observations
and, possibly, an initial state y0 , the predicted trajectory Z ? Rd?tend is,
zt = Azt?1 + W? (ot ) , tend ? t > 0 ,
(1)
where z0 = y0 , A ? Rd?d is a matrix describing linear movement dynamics and
W ? Rd?q is a weight matrix. ? (ot ) is a feature vector of the observed spike trains
at time t and is later replaced by a kernel operator (in the dual formulation to follow).
Thus, the state transition is a linear transformation of the previous state with the addition
of a non-linear effect of the observation.
Pt
Note that unfolding the recursion in Eq. (1) yields zt = At y0 + k=1 At?k W? (ok ) .
Assuming that A describes stable dynamics (the real parts of the eigenvalues of A are les
than 1), then the current prediction depends, in an exponentially decaying manner, on the
previous observations. We further assume that A is fixed and wish to learn W (we describe
our choice of A in Sec. 4). In addition, ot may also encompass a series of previous spike
counts in a window ending at time t (as is the case in Sec. 4). Also, note that this model (in
its non-kernelized version) has an algebraic form which is similar to the Kalman filter (to
which we compare our results later).
Primal Learning Problem: The optimization problem presented here is identical to the
standard SVR learning problem (see, for example [12]) with the exception that zit is defined
as in Eq. (1) while in standard SVR, zt = W? (ot
) (i.e. without the linear dynamics).
m
Given a training set of fully observed trials Yi , Oi i=1 we define the learning problem
to be
i
min
W
m tX
d
end X
X
i
1
2
zt ? yti .
kWk + c
s
s ?
2
i=1 t=1 s=1
(2)
P
2
2
Where kWk = a,b (W)ab (is the Forbenius norm). The second term is a sum of training
errors (in all trials, times and movement dimensions). | ? |? is the ? insensitive loss and is
defined as |v|? = max {0, |v| ? ?}. The first term is a regularization term that promotes
small weights and c is a fixed constant providing a tradeoff between the regularization
term and the training error. Note that to compensate for different units and scales of the
movement dimensions one could either define a different ?s and cs for each dimension of
the movement or, conversely, scale the sth movement dimension. The tracking method,
combined with the optimization specified here, defines the complete algorithm. We name
this method the Discriminative Dynamic Tracker or DDT in short.
A Dual Solution: The derivation of the dual of the learning problem defined in Eq. (2)
is rather mundane (e.g. [12]) and is thus omitted. Briefly, we replace the ?-loss with pairs
of slack variables. We then write a Lagrangian of the primal problem and replace zit with
its (less-standard) definition. We then differentiate the Lagrangian with respect to the slack
variables and W and obtain a dual optimization problem. We present the dual dual problem
in a top-down manner, starting with the general form and finishing with a kernel definition.
The form of the dual is
T
T
T
max? ? 12 (?? ? ?) G (?? ? ?) + (?? ? ?) y ? (?? + ?)
?,?
s.t. ?, ?? ? [0, c]
.
(3)
Note that the above expression conforms to the dual form of SVR. Let ` equal the size of the
movement space (d), multiplied by the total number of time steps in all the training trajectories. ?, ?? ? R` are vectors of Lagrange multipliers, y ? R` is a column concatenation of
T T
1 T
m
all the training set movement trajectories y1 ? ? ? ytm
, = [?, . . . , ?]T ? R`
end
and G ? R`?` is a Gram matrix (vT denotes transposition). One obvious difference between our setting and the standard SVR lies within the size of the vectors and Gram matrix.
In addition, a major difference is the definition of G. We define G here in a hierarchical
manner. Let i, j ? {1, . . . , m} be trajectory (trial) indexes. G is built from blocks indexed
by Gij , which are in turn made from basic blocks, indexed by Kij
tq as follows
?
?
?
?
ij
K11
???
Kij
G11 ? ? ? G1m
1tj
?
?
..
..
?
?
..
..
ij
..
?,
G = ? ...
? , G =?
.
.
.
.
.
?
?
ij
ij
Kti 1 ? ? ? Kti tj
Gm1 ? ? ? Gmm
end
end end
where block Gij refers to a pair of trials (i and j). Finally Each basic block, Kij
tq refers to a
j
i
pair of time steps t and q in trajectories i and j respectively. tend , tend are the time lengths
of trials i and j. Basic blocks are defined as
q
t X
X
ij
T
=
Kij
At?r krs
Aq?s ,
(4)
tq
r=1 s=1
ij
= k oir , ojs is a (freely chosen) basic kernel between the two neural observawhere krs
tions oir and ojs at times r and s in trials i and j respectively. For an explanation of kernel
operators we refer the
readerto [14] and mention that the kernel operator can be viewed
as computing ? oir ? ? ojs where ? is a fixed mapping to some inner product space.
The choice of kernel (being the choice of feature space) reflects a modeling decision that
specifies how similarities
P between neural patterns are measured. The resulting dual form
of the tracker is zt = k ?k Gtk where Gt is the Gram matrix row of the new example.
It is therefore clear from Eq. (4) that the linear dynamic characteristics of DDT results in
a Gram matrix whose entries depend on previous observations. This dependency is exponentially decaying as the time difference between events in the trajectories grow. Note
that solution of the dual optimization problem in Eq. (3) can be calculated by any standard quadratic programming optimization tool. Also, note that direct calculation of G is
inefficient. We describe an efficient method in the sequel.
Efficient Calculation of the Gram Matrix Simple, straight-forward calculation of the
Gram matrix is time consuming. To illustrate this, suppose each trial is of length tiend = n,
then calculation of each basic block would take ?(n2 ) summation steps. We now describe
a procedure based on dynamic-programming method for calculating the Gram matrix in a
constant number of operations for each basic block.
Omitting the indexing over trials to ease notation, we are interested in calculating the basic
Pt
block Ktq . First, define Btq = k=1 kkq At?k . the basic block Ktq can be recursively
calculated in three different ways:
Ktq = Kt(q?1) AT + Btq
(5)
Ktq
= AK(t?1)q + (Bqt )
T
T
(6)
T
Ktq = AK(t?1)(q?1) A + (Bqt ) + Btq ? ktq .
(7)
Thus, by adding Eq. (5) to Eq. (6) and subtracting Eq. (7) we get
Ktq = AK(t?1)q + Kt(q?1) AT ? AK(t?1)(q?1) AT + ktq I .
Btq (and the entailed summation) is eliminated in exchange for a 2D dynamic program with
initial conditions: K11 = k11 I , K1q = K1(q?1) AT + k1q I , Kt1 = AK(t?1)1 + kt1 I.
Table 1: Mean R2 , MAE? & MSE (across datasets, folds, hands and directions) for each algorithm.
Algorithm
Kalman filter
DDT-linear
SVR-Spikernel
DDT-Spikernal
pos.
0.64
0.59
0.61
0.73
R2
vel.
0.58
0.49
0.64
0.67
accl.
0.30
0.17
0.37
0.40
pos.
0.40
0.63
0.44
0.37
MAE?
vel.
0.15
0.41
0.14
0.14
accl.
0.37
0.58
0.34
0.34
pos.
0.78
0.97
0.76
0.50
MSE
vel. accl.
0.27 1.16
0.50 1.23
0.20 0.98
0.16 0.91
DDT?Spikernel, R2 Scores
1
0.8
0.6
0.4
left hand, X dir.
left hand, Y dir.
0.2
right hand, X dir.
right hand, Y dir.
0
0
0.2
0.4
0.6
0.8
Kalman filter, R2 Scores
1
0
0.2
0.4
0.6
0.8
DDT?linear, R2 Scores
1
0
0.2
0.4
0.6
0.8
SVR?Spikernel, R2 Scores
Figure 1: Correlation coefficients (R2 , of predicted and observed hand positions) comparisons of
the DDT-Spikernel versus the Kalman filter (left), DDT-linear (center) and SVR-Spikernel (right).
Each data point is the R2 values obtained by the DDT-Spikernel and by another method in one fold
of one of the datasets for one of the two axes of movement (circle / square) and one of the hands
(filled/non-filled). Results above the diagonals are cases were the DDT-Spikernel outperformes.
Suggested Optimization Method. One possible way to solve the optimization problem
(essentially, a modification of the method described in [4] for classification) is to sequentially solve a reduced problem with respect to a single constraint at a time. Define:
X ?
X ?
?i =
?j ? ?j Gij ? yi ? min
?j ? ?j Gij ? yi .
?
?i ,?i ?[0,c]
j
j
?
?
Then ?i is the amount of ?-insensitive error that can be corrected for example i by keeping
(?)
(?)
all ?j6=i constant and changing ?i . Optimality is reached by iteratively choosing the
(?)
example with the largest ?i and changing its ?i
error for this example.
4
within the [0, c] limits to minimize the
Experimental Setting
The data used in this work was recorded from the primary motor cortex of a Rhesus
(Macaca Mulatta) monkey (?4.5 kg). The monkey sat in a dark chamber, and up to 8
electrodes were introduced into MI area of each hemisphere. The electrode signals were
amplified, filtered and sorted. The data used in this report was recorded on 8 different days
and includes hand positions, sampled at 500Hz, spike times of single units (isolated by signal fit to a series of windows) and of multi units (detection by threshold crossing) sampled
at 1ms precision. The monkey used two planar-movement manipulanda to control 2 cursors on the screen to perform a center-out reaching task. Each trial began when the monkey
centered both cursors on a central circle. Either cursor could turn green, indicating the hand
to be used in the trial. Then, one of eight targets appeared (?go signal?), the center circle
disappeared and the monkey had to move and reach the target to receive liquid reward. The
number of multi-unit channels ranged from 5 to 15, the number of single units was 20-27
and the average total was 34 units per dataset. The average spike rate per channel was 8.2
spikes/sec. More information on the recordings can be found in [9].
1
DDT (Spikernel)
DDT (Spikernel)
75%
88.1%
100%
100%
Kalman Filter
78.12%
80.0%
96.3%
Kalman Filter
DDT (Linear)
91.88%
SVR (Spikernel)
98.7%
86.3%
95.6%
86.8%
62.5%
DDT (Linear)
78.7%
87.5%
SVR (Spikernel)
99.4%
63.75%
SVR (Spikernel)
DDT (Spikernel)
Kalman Filter
84.4%
DDT (Linear)
Figure 2: Comparison of R2 -performance between algorithms. Each algorithm is represented by a
vertex. The weight of an edge between two algorithms is the fraction of tests in which the algorithm
on top achieves higher R2 score than the other. A bold edge indicates a fraction higher than 95%.
Graphs from left to right are for position, velocity, and acceleration respectively.
The results that we present here refer to prediction of instantaneous hand movements during
the period from ?Go Signal? to ?Target Reach? times of both hands in successful trials.
Note that some of the trials required movement of the left hand while keeping the right
hand steady and vise versa. Therefore, although we considered only movement periods
of the trials, we had to predict both movement and non-movement for each hand. The
cumulative time length of all the datasets was about 67 minutes. Since the correlation
between the movements of the two hands tend to zero - we predicted movement for each
T
hand separately, choosing the movement space to be [x, y, vx , vy , ax , ay ] for each of the
T
hands (preliminary results using only [x, y, vx , vy ] were less accurate).
We preprocessed the spike trains into spike counts in a running windows of 100ms (choice
of window size is based on previous experience [11]). Hand position, velocity and acceleration were calculated using the 500Hz recordings. Both spike counts and hand movement
were then sampled at steps of 100ms (preliminary results
with
step size 50ms were negligibly different for all algorithms). A labeled example yti , oit for time t in trial i consisted
of the previous 10 bins of population spike counts and the state, as a 6D vector for the left
or right hand. Two such consecutive examples would than have 9 time bins of spike count
overlap. For example, the number of cortical units q in the first dataset was 43 (27 single
and 16 multiple) and the total length of all the trials that were used in that dataset is 529
seconds. Hence in that session there are 5290 consecutive examples where each is a 43?10
matrix of spike counts along with two 6D vectors of end point movement.
In order to run our algorithm we had to choose base kernels, their parameters, A and c (and
?, to be introduced below). We used the Spikernel [11], a kernel designed to be used with
spike rate patterns, and the simple dot product (i.e. linear regression). Kernel parmeters and
c were chosen (and subsequently held fixed) by 5 fold cross validation over half of the first
dataset only. We compared DDT with the Spikernel and with the linear kernel to standard
SVR using the Spikernel and the Kalman filter. We also obtained tracking results using
both DDT and SVR with the standard exponential kernel. These results were slightly less
accurate on average than with the Spikernel and are therefore omitted here. The Kalman
filter was learned assuming the standard state space model (yt = Ayt?1 + ? , ot =
Hyt +?, where ?, ? are white Gaussian noise with appropriate correlation matrices) such as
in [16]. y belonged to the same 6D state space as described earlier. To ease the comparison
- the same matrix A that was learned for the Kalman filter was used in our algorithm
(though we show that it is not optimal for DDT), multiplied by a scaling parameter ?. This
parameter was selected to produce best position results on the training set. The selected ?
value is 0.8.
The figures that we show in Sec. 5 are of test results in 5 fold cross validation on the rest
of the data. Each of the 8 remaining datasets was divided into 5 folds. 4/5 were used for
MSE
X
# Support
Position
MAE
14K
Y
position
2
R
Acceleration
8K
6K
?
?
?
?
Figure 3: Effect of ? on R2 , MAE? ,MSE and
number of support vectors.
velocity
10K
Actual
DDT?Spikernel
SVR?Spikernel
acceleration
Velocity
12K
Figure 4: Sample of tracking with the DDTSpikernel and the SVR-Spikernel.
training (with the parameters obtained previously and the remaining 1/5 as test set). This
process was repeated 5 times for each hand. Altogether we had 8sets ? 5folds ? 2hands = 80
folds.
5
Results
We begin by showing average results across all datasets, folds, hands and X/Y directions for
the four algorithms that are compared. Table. 1 shows mean Correlation Coefficients (R2 ,
between recorded and predicted movement values), Mean ? insensitive Absolute Errors
(MAE? ) and Mean Square Errors (MSE). R2 is a standard performance measure, MAE ? is
the error minimized by DDT (subject to the regularization term) and MSE is minimized by
the Kalman filter. Under all the above measures the DDT-Spikernel outperforms the rest
with the SVR-Spikernel and the Kalman Filter alternating in second place.
To understand whether the performance differences are significant we look at the distribution of position (X and Y) R2 values at each of the separate tests (160 altogether). Figure 1
shows scatter plots of R2 results for position predictions. Each plot compares the DDTSpikernel (on the Y axis) with one of the other three algorithms (on the X axes). It is
clear that in spite large differences in accuracy across datasets, the algorithm pairs achieve
similar success with the DDT-Spikernel achieving a better R2 score in almost all cases.
To summarize the significance of R2 differences we computed the number of tests in which
one algorithm achieved a higher R2 value than another algorithm (for all pairs, in each of
the position, velocity and acceleration categories). The results of this tournament between
the algorithms are presented in Figure 2 as winning percentages. The graphs produce a
ranking of the algorithms and the percentages are the significances of the ranking between
pairs. The DDT-Spikernel is significantly better then the rest in tracking position.
The matrix A in use is not optimal for our algorithm. The choice of ? scales its effect. When
? = 0 we get the standard SVR algorithm (without state dynamics). To illustrate the effect
of ? we present in Figure 3 the mean (over 5 folds, X/Y direction and hand) R2 results on
the first dataset as a function of ?. It is clear that the value chosen to minimize position error
is not optimal for minimizing velocity and acceleration errors. Another important effect of
? is the number of the support patterns in the learned model, which drops considerably
(by about one third) when the effect of the dynamics is increased. This means that more
training points fall strictly within the ?-tube in training, suggesting that the kernel which
tacitly results from the dynamical model is better suited for the problem. Lastly, we show a
sample of test tracking results for the DDT-Spikernel and SVR-Spikernel in Figure 4. Note
that the acceleration values are not smooth and are, therefore, least aided by the dynamics of
the model. However, adding acceleration to the model improves the prediction of position.
6
Conclusion
We described and reported experiments with a dynamical system that combines a linear
state mapping with a nonlinear observation-to-state mapping. The estimation of the system?s parameters is transformed to a dual representation and yields a novel kernel for temporal modelling. When a linear kernel is used, the DDT system has a similar form to the
Kalman filter as t ? ?. However, the system?s parameters are set so as to minimize the
regularized ?-insensitive `1 loss between state trajectories. DDT also bares similarity to
SVR, which employs the same loss yet without the state dynamics. Our experiments indicate that by combining a kernel-induced feature space, linear state dynamics, and using a
robust loss we are able to leverage the trajectory prediction accuracy and outperform common approaches. Our next step toward an accurate brain-machine interface for predicting
hand movements is the development of a learning procedure for the state dynamic mapping
A and further developments of neurally motivated and compact representations.
Acknowledgments This study was partly supported by a center of excellence grant (8006/00)
administered by the ISF, BMBF-DIP, by the U.S. Israel BSF and by the IST Programme of the European Community, under the PASCAL Network of Excellence, IST-2002-506778. L.S. is supported
by a Horowitz fellowship.
References
[1] A. E. Brockwell, A. L. Rojas, and R. E. Kass. Recursive bayesian decoding of motor cortical
signals by particle filtering. Journal of Neurophysiology, 91:1899?1907, 2004.
[2] E. N. Brown, L. M. Frank, D. Tang, M. C. Quirk, and M. A. Wilson. A statistical paradigm for
neural spike train decoding applied to position prediction from ensemble firing patterns of rat
hippocampal place cells. Journal of Neuroscience, 18(7411?7425), 1998.
[3] J. M. Carmena, M. A. Lebedev, R. E. Crist, J. E. O?Doherty, D. M. Santucci, D. F. Dimitrov,
P. G. Patil, C. S. Henriques, and M. A. L. Nicolelis. Learning to control a brain-machine
interface for reaching and grasping by primates. PLOS Biology, 1(2):001?016, 2003.
[4] K. Crammer and Y. Singer. On the algorithmic implementation of multiclass kernel-based
vector machines. Jornal of Machine Learning Research, 2:265?292, 2001.
[5] J. Eichhorn, A. Tolias, A. Zien, M. Kuss, C. E. Rasmussen, J. Weston, N. Logothetis, and
B. Sch?olkopf. Prediction on spike data using kernel algorithms. In NIPS 16. MIT Press, 2004.
[6] A. P. Georgopoulos, J. Kalaska, and J. Massey. Spatial coding of movements: A hypothesis
concerning the coding of movement direction by motor cortical populations. Experimental
Brain Research (Supp), 7:327?336, 1983.
[7] R. E. Isaacs, D. J. Weber, and A. B. Schwartz. Work toward real-time control of a cortical
neural prothesis. IEEE Trans Rehabil Eng, 8(196?198), 2000.
[8] C. Mehring, J. Rickert, E. Vaadia, S. C. de Oliveira, A. Aertsen, and S. Rotter. Inference of
hand movements from local field potentials in monkey motor cortex. Nature Neur., 6(12), 2003.
[9] R. Paz, T. Boraud, C. Natan, H. Bergman, and E. Vaadia. Preparatory activity in motor cortex
reflects learning of local visuomotor skills. Nature Neur., 6(8):882?890, August 2003.
[10] M. D. Serruya, N. G. Hatsopoulos, L. Paninski, M. R. Fellows, and J. P. Donoghue. Instant
neural control of a movement signal. Nature, 416:141?142, March 2002.
[11] L. Shpigelman, Y. Singer, R. Paz, and E. Vaadia. Spikernels: Embedding spiking neurons in
inner product spaces. In NIPS 15, Cambridge, MA, 2003. MIT Press.
[12] A. Smola and B. Scholkop. A tutorial on support vector regressio. In NeuroCOLT2 Technical
Report, 1998.
[13] S. I. H. Tillery, D. M. Taylor, and A. B. Schwartz. Training in cortical control of neuroprosthetic devices improves signal extraction from small neuronal ensembles. Reviews in the Neurosciences, 14:107?119, 2003.
[14] V. Vapnik. The Nature of Statistical Learning Theory. Springer, N.Y., 1995.
[15] J. Wessberg, C. R. Stambaugh, J. D. Kralik, P. D. Beck, M. Laubach, J. K. Chapin, J. Kim,
J. Biggs, M. A. Srinivasan, and M. A. Nicolelis. Real-time prediction of hand trajectory by
ensembles of cortical neurons in primates. Nature, 408(16), November 2000.
[16] W. Wu, M. J. Black, Y. Gao, E. Bienenstock, M. Serruya, and J. P. Donoghue. Inferring hand
motion from multi-cell recordings in motor cortex using a kalman filter. In SAB02, pages 66?73,
Edinburgh, Scotland (UK), 2002.
| 2644 |@word neurophysiology:1 trial:19 version:1 briefly:1 norm:1 open:2 rhesus:1 eng:1 mention:1 recursively:1 initial:2 contains:1 series:5 score:6 liquid:1 tuned:2 outperforms:1 current:2 comparing:1 ka:1 scatter:1 yet:1 additive:1 partition:1 eichhorn:1 motor:8 designed:1 plot:2 drop:1 generative:2 half:1 devising:1 selected:2 device:1 wessberg:1 scotland:1 short:1 filtered:1 transposition:1 characterization:1 location:2 along:2 direct:1 become:1 consists:1 combine:1 shpigelman:1 manner:3 introduce:1 excellence:2 preparatory:1 multi:3 brain:5 little:1 actual:1 inappropriate:1 window:4 provided:1 begin:1 underlying:1 notation:1 chapin:1 k1q:2 israel:2 what:1 kg:1 monkey:7 transformation:1 temporal:2 fellow:1 every:1 oir:3 gm1:1 schwartz:2 control:5 unit:11 medical:1 grant:1 uk:1 engineering:1 local:2 limit:1 ak:5 firing:5 crist:1 tournament:1 black:1 conversely:1 ease:2 jornal:1 acknowledgment:1 recursive:1 block:9 implement:1 procedure:2 area:1 physiology:1 significantly:1 refers:2 spite:1 svr:20 get:2 operator:3 eilon:1 spikernels:1 lagrangian:2 center:5 yt:2 jerusalem:1 go:2 starting:1 formalized:1 bsf:1 enabled:1 population:7 handle:1 embedding:1 variation:1 construction:1 target:5 pt:2 suppose:1 logothetis:1 programming:2 hypothesis:1 bergman:1 velocity:9 crossing:1 expensive:1 approximated:1 labeled:1 observed:4 negligibly:1 grasping:1 movement:50 plo:1 hatsopoulos:1 rq:2 pd:1 reward:1 dynamic:19 tacitly:1 depend:1 solving:1 basis:1 biggs:1 po:3 represented:1 tx:1 derivation:1 train:7 describe:5 visuomotor:1 choosing:2 whose:1 encoded:1 solve:2 azt:1 itself:1 differentiate:1 advantage:1 sequence:1 eigenvalue:1 vaadia:3 subtracting:1 product:3 loop:3 combining:1 rapidly:1 brockwell:1 achieve:2 amplified:1 tillery:1 macaca:1 olkopf:1 carmena:1 electrode:2 requirement:1 produce:2 disappeared:1 object:1 tions:1 illustrate:2 ac:1 quirk:1 measured:2 ij:6 school:2 eq:8 zit:3 c:2 predicted:6 indicate:1 direction:8 undertaking:1 filter:16 subsequently:1 centered:1 vx:2 bin:4 exchange:1 singer1:1 preliminary:2 summation:2 strictly:1 stambaugh:1 confine:1 tracker:2 considered:1 mapping:9 predict:2 algorithmic:1 major:1 vary:1 early:1 consecutive:3 omitted:2 achieves:1 estimation:1 largest:1 tool:1 weighted:1 reflects:2 unfolding:1 mit:2 gaussian:2 reaching:4 rather:1 eating:1 wilson:1 ax:3 focus:1 finishing:1 prevalent:1 indicates:1 modelling:1 greatly:1 contrast:1 kim:1 sense:1 inference:2 typically:4 kernelized:1 bienenstock:1 transformed:1 interested:2 overall:1 dual:12 classification:1 pascal:1 development:4 plan:1 spatial:2 constrained:1 art:1 equal:1 aware:1 field:1 having:1 extraction:1 eliminated:1 identical:1 biology:1 koby:1 look:1 lavi:1 future:1 minimized:2 report:2 employ:3 parmeters:1 beck:1 replaced:1 intended:2 ourselves:1 consisting:1 tq:3 attempt:1 ab:1 detection:1 introduces:1 entailed:1 primal:2 tj:2 held:1 accurate:5 kt:2 edge:2 hyt:1 experience:1 conforms:1 indexed:3 filled:2 euclidean:1 loosely:1 taylor:1 circle:3 causal:2 isolated:1 kij:4 column:1 modeling:4 earlier:1 increased:1 vertex:1 entry:1 successful:1 paz:2 reported:1 dependency:2 ojs:3 dir:4 considerably:1 combined:1 huji:1 interdisciplinary:1 sequel:2 probabilistic:1 off:1 decoding:3 lebedev:1 central:1 recorded:4 tube:1 choose:1 possibly:1 horowitz:1 inefficient:1 supp:1 suggesting:2 account:1 potential:1 de:1 sec:10 bold:1 includes:1 coefficient:2 coding:2 satisfy:1 bqt:2 ranking:2 depends:1 later:2 lab:1 closed:2 kwk:2 reached:1 decaying:2 ytm:1 correspondance:1 il:1 oi:3 square:2 accomplishing:1 accuracy:2 characteristic:1 minimize:3 ensemble:3 yield:3 bayesian:1 trajectory:17 straight:1 j6:1 kuss:1 reach:2 email:1 definition:3 ktq:8 obvious:1 isaac:1 naturally:1 associated:1 mi:1 static:2 sampled:3 dataset:5 improves:3 ok:1 higher:3 day:2 follow:1 reflected:1 planar:1 formulation:1 vel:3 though:1 spikernel:26 furthermore:1 smola:1 lastly:1 correlation:4 hand:39 nonlinear:2 defines:1 disabled:1 name:1 effect:6 omitting:1 ranged:1 multiplier:1 consisted:1 brown:1 regularization:3 hence:1 alternating:1 iteratively:1 white:1 during:3 steady:1 rat:1 m:5 prominent:1 hippocampal:1 ay:1 complete:1 doherty:1 performs:1 dedicated:1 interface:3 motion:1 weber:1 instantaneous:3 novel:1 began:1 common:1 specialized:1 mulatta:1 spiking:1 exponentially:2 insensitive:4 mae:6 isf:1 refer:2 significant:1 versa:1 cambridge:1 tuning:1 rd:3 session:2 particle:1 aq:1 had:4 dot:1 robot:1 stable:1 cortex:5 behaving:1 similarity:2 gt:1 base:1 hadassah:1 recent:1 showed:1 hemisphere:1 scenario:1 success:2 rotter:1 life:1 vt:1 yi:5 devise:2 freely:1 paradigm:1 period:2 signal:7 zien:1 neurally:2 encompass:1 multiple:1 stem:1 smooth:6 technical:1 calculation:5 cross:2 compensate:1 kalaska:1 divided:2 concerning:1 promotes:1 prediction:8 regression:6 basic:9 patient:1 essentially:1 kernel:27 tailored:2 serruya:2 achieved:1 cell:3 receive:1 addition:3 fellowship:1 separately:1 grow:1 dimitrov:1 sch:1 ot:6 rest:3 subject:3 induced:2 recording:4 tend:6 hz:2 incorporates:1 leverage:1 independence:1 fit:1 inner:2 tradeoff:1 multiclass:1 donoghue:2 oit:2 administered:1 whether:1 expression:1 motivated:1 algebraic:1 constitute:1 oi1:1 clear:3 involve:1 amount:1 dark:1 oliveira:1 category:1 reduced:1 specifies:1 outperform:1 percentage:2 vy:2 tutorial:1 estimated:1 neuroscience:2 per:3 track:1 ddt:27 write:1 srinivasan:1 ist:2 four:1 threshold:1 achieving:1 changing:2 preprocessed:1 gmm:1 kkq:1 massey:1 graph:2 fraction:2 sum:1 run:1 angle:1 place:2 almost:1 reader:1 wu:1 decision:1 scaling:1 mundane:1 fold:10 quadratic:1 activity:8 constraint:1 georgopoulos:1 underscoring:1 min:2 optimality:1 relatively:2 designated:1 neur:2 combination:1 march:1 describes:1 across:3 reconstructing:1 y0:3 sth:1 slightly:1 making:1 modification:1 primate:2 gradually:1 indexing:1 computationally:1 previously:1 turn:3 count:11 describing:2 slack:2 singer:2 merit:1 end:7 operation:1 multiplied:2 apply:1 eight:1 hierarchical:1 appropriate:1 chamber:1 altogether:2 original:1 top:2 denotes:1 running:1 remaining:2 patil:1 instant:1 calculating:2 yoram:1 kt1:2 k1:1 move:2 spike:24 neurocolt2:1 costly:1 primary:1 diagonal:1 aertsen:1 lends:1 distance:2 separate:2 concatenation:1 toward:2 assuming:2 kalman:16 length:4 modeled:2 index:1 manipulanda:1 providing:2 minimizing:2 hebrew:1 laubach:1 frank:1 kralik:1 implementation:2 zt:5 perform:1 neuron:3 observation:5 datasets:6 november:1 displayed:1 situation:1 y1:1 august:1 community:1 g1m:1 natan:1 introduced:2 pair:7 required:2 specified:1 rickert:1 learned:4 nip:2 trans:1 able:1 suggested:1 dynamical:6 pattern:4 below:1 appeared:1 belonged:1 summarize:1 program:1 built:1 max:2 green:1 explanation:1 event:1 overlap:1 natural:1 nicolelis:2 regularized:1 predicting:1 santucci:1 recursion:1 arm:1 axis:1 ayt:1 review:1 understanding:1 loss:6 fully:1 rationale:1 generation:1 filtering:3 versus:1 validation:2 kti:2 imposes:1 row:1 summary:1 supported:2 free:1 keeping:2 rasmussen:1 henriques:1 understand:1 fall:1 absolute:1 edinburgh:1 dip:1 feedback:1 dimension:4 cortical:9 transition:2 ending:2 gram:7 calculated:3 cumulative:1 forward:1 made:1 neuroprosthetic:1 programme:1 compact:1 skill:1 preferred:2 sequentially:1 sat:1 consuming:1 discriminative:3 tolias:1 mehring:1 table:2 nature:6 learn:2 robust:2 channel:2 mse:6 european:1 domain:1 significance:2 bmi:4 noise:3 n2:1 gtk:1 repeated:2 neuronal:2 screen:1 bmbf:1 precision:2 position:17 inferring:1 wish:3 exponential:1 winning:1 lie:1 third:1 tang:1 z0:1 down:1 minute:1 showing:1 g11:1 r2:19 k11:3 vapnik:1 adding:2 effectively:1 kr:2 cursor:4 easier:1 suited:2 vise:1 paninski:1 gao:1 forbenius:1 lagrange:1 hitting:1 tracking:10 springer:1 corresponds:1 ma:1 weston:1 viewed:2 goal:1 sorted:1 acceleration:12 rojas:1 towards:1 rehabil:1 replace:2 yti:5 hard:1 aided:1 corrected:1 shpigi:1 total:3 gij:4 forgo:1 experimental:3 partly:1 exception:1 formally:1 indicating:1 support:6 modulated:1 crammer:1 dept:1 |
1,811 | 2,645 | Confidence Intervals for the Area under the
ROC Curve
Corinna Cortes
Google Research
1440 Broadway
New York, NY 10018
[email protected]
Mehryar Mohri
Courant Institute, NYU
719 Broadway
New York, NY 10003
[email protected]
Abstract
In many applications, good ranking is a highly desirable performance for
a classifier. The criterion commonly used to measure the ranking quality
of a classification algorithm is the area under the ROC curve (AUC). To
report it properly, it is crucial to determine an interval of confidence for
its value. This paper provides confidence intervals for the AUC based
on a statistical and combinatorial analysis using only simple parameters
such as the error rate and the number of positive and negative examples.
The analysis is distribution-independent, it makes no assumption about
the distribution of the scores of negative or positive examples. The results
are of practical use and can be viewed as the equivalent for AUC of the
standard confidence intervals given in the case of the error rate. They
are compared with previous approaches in several standard classification
tasks demonstrating the benefits of our analysis.
1
Motivation
In many machine learning applications, the ranking quality of a classifier is critical. For
example, the ordering of the list of relevant documents returned by a search engine or
a document classification system is essential. The criterion widely used to measure the
ranking quality of a classification algorithm is the area under an ROC curve (AUC). But, to
measure and report the AUC properly, it is crucial to determine an interval of confidence
for its value as it is customary for the error rate and other measures. It is also important to
make the computation of the confidence interval practical by relying only on a small and
simple number of parameters. In the case of the error rate, such intervals are often derived
from just the sample size N .
We present an extensive theoretical analysis of the AUC and show that a similar confidence
interval can be derived for its value using only simple parameters such as the error rate k/N ,
the number of positive examples m, and the number of negative examples n = N ? m.
Thus, our results extend to AUC the computation of confidence intervals from a small
number of readily available parameters.
Our analysis is distribution-independent in the sense that it makes no assumption about the
distribution of the scores of negative or positive examples. The use of the error rate helps
determine tight confidence intervals. This contrasts with existing approaches presented in
the statistical literature [11, 5, 2] which are based either on weak distribution-independent
assumptions resulting in too loose confidence intervals, or strong distribution-dependent
assumptions leading to tight but unsafe confidence intervals.
We show that our results are of practical use. We also compare them with previous approaches in several standard classification tasks demonstrating the benefits of our analysis.
Our results are also useful for testing the statistical significance of the difference of the
AUC values of two classifiers.
The paper is organized as follows. We first introduce the definition of the AUC, its connection with the Wilcoxon-Mann-Whitney statistic (Section 2), and briefly review some
essential aspects of the existing literature related to the computation of confidence intervals
for the AUC. Our computation of the expected value and variance of the AUC for a fixed
error rate requires establishing several combinatorial identities. Section 4 presents some
existing identities and gives the proof of novel ones useful for the computation of the variance. Section 5 gives the reduced expressions for the expected value and variance of the
AUC for a fixed error rate. These can be efficiently computed and used to determine our
confidence intervals for the AUC (Section 6). Section 7 reports the result of the comparison
of our method with previous approaches, including empirical results for several standard
tasks.
2
Definition and Properties of the AUC
The Receiver Operating Characteristics (ROC) curves were originally introduced in signal
detection theory [6] in connection with the study of radio signals, and have been used
since then in many other applications, in particular for medical decision-making. Over the
last few years, they have found increased interest in the machine learning and data mining
communities for model evaluation and selection [14, 13, 7, 12, 16, 3]. The ROC curve for
a binary classification problem plots the true positive rate as a function of the false positive
rate. The points of the curve are obtained by sweeping the classification threshold from the
most positive classification value to the most negative. For a fully random classification,
the ROC curve is a straight line connecting the origin to (1, 1). Any improvement over
random classification results in an ROC curve at least partially above this straight line. The
AUC is defined as the area under the ROC curve.
Consider a binary classification task with m positive examples and n negative examples.
Let C be a fixed classifier that outputs a strictly ordered list for these examples. Let
x1 , . . . , xm be the output of C on the positive examples and y1 , . . . , yn its output on the
negative examples and denote by 1X the indicator function of a set X. Then, the AUC, A,
associated to C is given by:
Pm Pn
i=1
j=1 1xi >yj
(1)
A=
mn
which is the value of the Wilcoxon-Mann-Whitney statistic [10]. Thus, the AUC is closely
related to the ranking quality of the classification. It can be viewed as a measure based on
pairwise comparisons between classifications of the two classes. It is an estimate of the
probability Pxy that the classifier ranks a randomly chosen positive example higher than
a negative example. With a perfect ranking, all positive examples are ranked higher than
the negative ones and A = 1. Any deviation from this ranking decreases the AUC, and the
expected AUC value for a random ranking is 0.5.
3
Overview of Related Work
This section briefly describes some previous distribution-dependent approaches presented
in the statistical literature to derive confidence intervals for the AUC and compares them
to our method. The starting point for these analyses is a formula giving the variance of the
AUC, A, for a fixed distribution of the scores Px of the positive examples and Py of the
negative examples [10, 1]:
2
?A
=
A(1 ? A) + (m ? 1)(Pxxy ? A2 ) + (n ? 1)(Pxyy ? A2 )
mn
(2)
where Pxxy is the probability that the classifier ranks two randomly chosen positive examples higher than a negative one, and Pxyy the probability that it ranks two randomly chosen
negative examples lower than a positive one. To compute the variance exactly using Equation 2, the distributions Px and Py must be known.
Hanley and McNeil [10] argue in favor of exponential distributions, loosely claiming that
this upper-bounds the variance of normal distributions with various means and ratios of
A
2A2
variances. They show that for exponential distributions Pxxy = 2?A
and Pxyy = 1+A
.
The resulting confidence intervals are of course relatively tight, but their validity is questionable since they are based on a strong assumption about the distributions of the positive
and negative scores that may not hold in many cases.
An alternative considered by several authors to the exact computation of the variance is to
determine instead the maximum of the variance over all possible continuous distributions
with the same expected value of the AUC. For all such distributions, one can fix m and
n and compute the expected AUC and its variance. The maximum variance is denoted by
2
?max
and is given by [5, 2]:
2
?max
=
A(1 ? A)
1
?
min {m, n}
4 min {m, n}
(3)
Unfortunately, this often yields loose confidence intervals of limited practical use.
Our approach for computing the mean and variance of the AUC is distribution-independent
and inspired by the machine learning literature where analyses typically center on the error
rate. We require only that the error rate be measured and compute the mean and variance of
the AUC over all distributions Px and Py that maintain the same error rate. Our approach
is in line with that of [5, 2] but it crucially avoids considering the maximum of the variance. We show that it is possible to compute directly the mean and variance of the AUC
assigning equal weight to all the possible distributions. Of course, one could argue that not
all distributions Px and Py are equally probable, but since these distributions are highly
problem-dependent, we find it risky to make any general assumption on the distributions
and thereby limit the validity of our results. Our approach is further justified empirically
by the experiments reported in the last section.
4
Combinatorial Analysis
The analysis of the statistical properties of the AUC given a fixed error rate requires various
combinatorial calculations. This section describes several of the combinatorial identities
that are used in our computation of the confidence intervals. For all q ? 0, let Xq (k, m, n)
be defined by:
0
k
X
M
M
Xq (k, m, n) =
xq
(4)
x
x0
x=0
where M = m ? (k ? x) + x, M 0 = n + (k ? x) ? x, and x0 = k ? x. In previous work,
we derived the following two identities which we used to compute the expected value of
the AUC [4]:
k
k
X
X
n+m+1
(k ? x)(m ? n) + k n + m + 1
X1 (k, m, n) =
X0 (k, m, n) =
x
2
x
x=0
x=0
To simplify the expression of the variance of the AUC, we need to compute X2 (k, m, n).
Proposition 1 Let k, m, n be non-negative integers such that k ? min{m, n}, then:
k
X
m+n+1
X2 (k, m, n) =
P2 (k, m, n, x)
x
x=0
(5)
where P2 is the following 4th-degree polynomial: P2 (k, m, n, x) = (k ? x)/12(?2x3 +
2x2 (2m ? n + 2k ? 4) + x(?3m2 + 3nm + 3m ? 5km ? 2k 2 + 2 + k + nk + 6n) +
(3(k ? 1)m2 ? 3nm(k ? 1) + 6km + 5m + k 2 m + 8n + 8 ? 9nk + 3k + k 2 + k 2 n)).
Proof.
5
The proof of the proposition is left to a longer version of this paper.
Expectation and Variance of the AUC
This section presents the expression of the expectation and variance of the AUC for a fixed
error rate k/N assuming that all classifications or rankings with k errors are equiprobable.
For a given classification, there may be x, 0 ? x ? k, false positive examples. Since the
number of errors is fixed, there are x0 = k ? x false negative examples. The expression Xq
discussed in the previous section represents the q-th moment of x over all classifications
with exactly k errors. In previous work, we gave the exact expression of the expectation of
the AUC for a fixed number of errors k:
Proposition 2 ([4]) Assume that a binary classification task with m positive examples
and n negative examples is given. Then, the expected value of the AUC, A, over all
classifications with k errors is given by:
Pk?1 m+n !
k
k
(n ? m)2 (m + n + 1)
x
E[A] = 1 ?
?
? Pk x=0 m+n+1
.
m+n
4mn
m+n
x=0
x
Note that the two sums in this expression cannot be further simplified since they are known
not to admit a closed form [9]. We also gave the expression of the variance of the AUC in
terms of the function F defined for all Y by:
Pk
M M0
x=0 x
x0 Y
F (Y ) = Pk
(6)
.
M M0
x=0
x
x0
The following proposition reproduces that result:
Proposition 3 ([4]) Assume that a binary classification task with m positive examples
and n negative examples is given. Then, the variance of the AUC A over all classifications with k errors is given by: ? 2 (A) = F ((1 ?
k?x
x
n+ m
)2 )
2
2
2
F ( mx +n(k?x) +(m(m+1)x+n(n+1)(k?x))?2x(k?x)(m+n+1)
).
12m2 n2
? F ((1 ?
k?x
x
n+ m
2
))2 +
Because of the products of binomial terms, the computation of the variance using this
expression is inefficient even for relatively small values of m and n. This expression can
however be reduced using the identities presented in the previous section which leads to
significantly more efficient computations that we have been using in all our experiments.
Corollary 1 ([4]) Assume that a binary classification task with m positive examples and n
negative examples is given. Then, the variance of the AUC A over all classifications with
((m+n?2)Z4 ?(2m?n+3k?10)Z3 )
k errors is given by: ? 2 (A) = (m+n+1)(m+n)(m+n?1)T72m
+
2 n2
(m+n+1)(m+n)T (m2 ?nm+3km?5m+2k2 ?nk+12?9k)Z2
48m2 n2
(m+n+1)Q1 Z1
kQ0
+
2
2
2
72m
n
144m n2 with:
Pk?i
m+n+1?i
Zi =
x=0
Pk
(
x=0
x
(m+n+1
)
x
)
?
(m+n+1)2 (m?n)4 Z12
16m2 n2
?
, T = 3((m ? n)2 + m + n) + 2, and:
Q0 = (m + n + 1)T k2 + ((?3n2 + 3mn + 3m + 1)T ? 12(3mn + m + n) ? 8)k + (?3m2 +
7m + 10n + 3nm + 10)T ? 4(3mn + m + n + 1)
Q1 = T k3 + 3(m ? 1)T k2 + ((?3n2 + 3mn ? 3m + 8)T ? 6(6mn + m + n))k + (?3m2 +
7(m + n) + 3mn)T ? 2(6mn + m + n)
Proof. The expression of the variance given in Proposition 3 requires the computation
of Xq (k, m, n), q = 0, 1, 2. Using the identities giving the expressions of X0 and X1 and
Proposition 1, which provides the expression of X2 , ? 2 (A) can be reduced to the expression
given by the corollary.
6
Theory and Analysis
Our estimate of the confidence interval for the AUC is based on a simple and natural assumption. The main idea for its computation is the following. Assume that a confidence
interval E = [e1 , e2 ] is given for the error rate of a classifier C over a sample S, with the
confidence level 1 ? . This interval may have have been derived from a binomial model of
C, which is a standard assumption for determining a confidence interval for the error rate,
or from any other model used to compute that interval. For a given error rate e ? E, or
equivalently for a given number of misclassifications, we can use the expectation and variance computed in the previous section and Chebyshev?s inequality to predict a confidence
interval Ae for the AUC at the confidence level 1 ? 0 . Since our equiprobable model for
the classifications is independent of the model used to compute the interval of confidence
for the error rate, we can use E and Ae , e ? E, to compute a confidence interval of the
AUC at the level (1 ? )(1 ? 0 ).
Theorem 1 Let C be a binary classifier and let S be a data sample of size N with m
positive examples and n negative examples, N = m + n. Let E = [e1 , e2 ] be a confidence
interval for the error rate of C over S at the confidence level 1 ? . Then, for any 0 ,
0 ? 0 ? 1, we can compute a confidence interval for the AUC value of the classifier C at
the confidence level (1 ? )(1 ? 0 ) that depends only on , 0 , m, n, and the interval E.
Proof. Let k1 = N e1 and k2 = N e2 be the number of errors associated to the error rates
e1 and e2 , and let IK be the interval IK = [k1 , k2 ]. For a fixed k ? IK , by Propositions
2 and Corollary 1, we can compute the exact value of the expectation E[Ak ] and variance
? 2 (Ak ) of the AUC Ak . Using Chebyshev?s inequality, for any k ? IK and any k > 0,
?(Ak )
P |Ak ? E[Ak ]| ? ?
? k
(7)
k
where E[Ak ] and ?(Ak ) are the expressions given in Propositions 2 and Corollary 1, which
depend only on k, m, and n. Let ?1 and ?2 be defined by:
?(Ak )
?(Ak )
?2 = max E[Ak ] + ?
(8)
?1 = min E[Ak ] ? ?
k?IK
k?IK
k
k
?1 and ?2 only depend on IK (i.e., on e1 and e2 ), and on k, m, and n. Let IA be the
confidence interval defined by IA = [?1 , ?2 ] and let k = 0 for all k ? IK . Using the
fact that the confidence interval E is independent of our equiprobability model for fixed-k
AUC values and the Bayes? rule:
X
P(A ? IA ) =
P (A ? IA | K = k)P (K = k)
(9)
k?R+
?
X
k?IK
P (A ? IA | K = k)P (K = k)
? (1 ? 0 )
X
k?IK
P (K = k) ? (1 ? 0 )(1 ? )
(10)
(11)
where we used the property of Eq. 7 and the definitions of the intervals IK and IA . Thus,
IA constitutes a confidence interval for the AUC value of C at the confidence level (1 ?
)(1 ? 0 ).
In practice, the confidence interval E is often determined as a result of the assumption that
C follows a binomial law. This leads to the following theorem.
.020
.035
Standard deviation
Standard deviation
.030
.015
.010
Max
Distribution?dependent
Distribution?independent
.005
.025
.020
.015
Max
Distribution?dependent
Distribution?independent
.010
.005
0.75
0.80
0.85
0.90
0.95
1.00
0.6
0.7
0.8
AUC
AUC
(a)
(b)
0.9
1.0
Figure 1: Comparison of the standard deviations for three different methods with: (a) m = n = 500;
(b) m = 400 and n = 200. The curves are obtained by computing the expected AUC and its standard
deviations for different values of the error rate using the maximum-variance approach (Eq. 3), our
distribution-independent method, and the distribution-dependent approach of Hanley [10].
Theorem 2 Let C be a binary classifier, let S be a data sample of size N with m positive
examples and n negative examples, N = m + n, and let k0 be the number of misclassifications of C on S. Assume that C follows a binomial law, then, for any , 0 ? ? 1, we can
compute a confidence interval of the AUC value of the classifier C at the confidence level
1 ? that depends only on , k0 , m, and n.
Proof.
Assume that C follows a binomial law with coefficient p. Then, Chebyshev?s
inequality yields:
1
p(1 ? p)
?
(12)
P(|C ? E[C]| ? ?) ?
2
N?
4N ? 2
1
1
Thus, E = [ kN0 ? ? ?
, k0 + ? ?
] forms a confidence interval for the
2 (1? 1?)N N
2 (1? 1?)N
?
error rate of C at the confidence level 1 ? . By Theorem
1, we can?compute for the
?
AUC value a confidence interval at the level (1 ? (1 ? 1 ? ))(1 ? (1 ? 1 ? )) = 1 ?
depending only on , m, n, and the interval E, i.e., k0 , N = m + n, and .
For large N , we can use the normal approximation of the binomial law to determine a finer
interval E. Indeed, for large N ,
?
(13)
P(|C ? E[C]| ? ?) ? 2?(2 N ?)
with ?(u) =
R?
u
2
/2
e?x
?
dx.
2?
Thus, E = [ kN0 ?
?
??1 ( 1? 21? ) k0
?
,N
2 N ?
confidence interval for the error rate at the confidence level
?
+
??1 ( 1? 21? )
?
]
2 N
is the
1 ? .
For simplicity, in the proof of Theorem 2, k was chosen to be a constant (k = 0 ) but, in
general, it can be another function of k leading to tighter confidence intervals. The results
presented in the next section were obtained with k = a0 exp((k ? k0 )2 /2a21 ), where a0
and a1 are constants selected so that the inequality 11 be verified.
7
Experiments and Comparisons
The analysis in the previous section provides a principled method for computing a confidence interval of the AUC value of a classier C at the confidence level 1 ? that depends
only on k, n and m. As already discussed, other expressions found in the statistical literature lead to either too loose or unsafely narrow confidence intervals based on questionable assumptions on the probability functions Px and Py [10, 15]. Figure 1 shows a
comparison of the standard deviations obtained using the maximum-approach (Eq. 3), the
distribution-dependent expression from [10], and our distribution-independent method for
NAME
m+n
n
m+n
AUC
k
m+n
?indep
?A
?dep
?max
368
700
303
1159
2473
201
0.63
0.67
0.54
0.17
0.10
0.37
0.70
0.63
0.87
0.85
0.84
0.85
0.24
0.26
0.13
0.05
0.03
0.13
0.0297
0.0277
0.0176
0.0177
0.0164
0.0271
0.0440
0.0330
0.0309
0.0161
0.0088
0.0463
0.0269
0.0215
0.0202
0.0176
0.0161
0.0306
0.0392
0.0317
0.0281
0.0253
0.0234
0.0417
pima
yeast
credit
internet-ads
page-blocks
ionosphere
Table 1: Accuracy and AUC values for AdaBoost [8] and estimated standard deviations for several
datasets from the UC Irvine repository. ?indep is a distribution-independent standard deviation obtained using our method (Theorem 2). ?A is given by Eq. (2) with the values of A, Pxxy , and Pxyy
derived from data. ?dep is the distribution-dependent standard deviation of Hanley [10], which is
based on assumptions that may not always hold. ?max is defined by Eq. (3). All results were obtained
on a randomly selected test set of size m + n.
various error rates. For m = n = 500, our distribution-independent method consistently
leads to tighter confidence intervals (Fig. 1 (a)). It also leads to tighter confidence intervals for AUC values more than .75 for the uneven distribution m = 400 and n = 200
(Fig. 1 (b)). For lower AUC values, the distribution-dependent approach produces tighter
intervals, but its underlying assumptions may not hold.
A different comparison was made using several datasets available from the UC Irvine repository (Table 1). The table shows that our estimates of the standard deviations (?indep ) are in
general close to or tighter than the distribution-dependent standard deviation ?dep of Hanley
[10]. This is despite we do not make any assumption about the distributions of positive
and negative examples. In contrast, Hanley?s method is based on specific assumptions
about these distributions. Plots of the actual ranking distribution demonstrate that these
assumptions are often violated however. Thus, the relatively good performance of Hanley?s approach on several data sets can be viewed as fortuitous and is not general. Our
distribution-independent method provides tight confidence intervals, in some cases tighter
than those derived from ?A , in particular because it exploits the information provided by
the error rate. Our analysis can also be used to determine if the AUC values produced by
two classifiers are statistically significant by checking if the AUC value of one falls within
the confidence interval of the other.
8
Conclusion
We presented principled techniques for computing useful confidence intervals for the AUC
from simple parameters: the error rate, and the negative and positive sample sizes. We
demonstrated the practicality of these confidence intervals by comparing them to previous
approaches in several tasks. We also derived the exact expression of the variance of the
AUC for a fixed k, which can be of interest in other analyses related to the AUC.
The Wilcoxon-Mann-Whitney statistic is a general measure of the quality of a ranking that
is an estimate of the probability that the classifier ranks a randomly chosen positive example higher than a negative example. One could argue that accuracy at the top or the
bottom of the ranking is of higher importance. This, however, contrarily to some belief,
is already captured to a certain degree by the definition of the Wilcoxon-Mann-Whitney
statistic which penalizes more errors at the top or the bottom of the ranking. It is however an interesting research problem to determine how to incorporate this bias in a stricter
way in the form of a score-specific weight in the ranking measure, a weighted WilcoxonMann-Whitney statistic, or how to compute the corresponding expected value and standard
deviation in a general way and design machine learning algorithms to optimize such a mea-
sure. A preliminary analysis suggests, however, that the calculation of the expectation and
the variance are likely to be extremely complex in that case. Finally, it could also be interesting but difficult to adapt our results to the distribution-dependent case and compare
them to those of [10].
Acknowledgments
We thank Rob Schapire for pointing out to us the problem of the statistical significance
of the AUC, Daryl Pregibon for the reference to [11], and Saharon Rosset for various
discussions about the topic of this paper.
References
[1] D. Bamber. The Area above the Ordinal Dominance Graph and the Area below the
Receiver Operating Characteristic Graph. Journal of Math. Psychology, 12, 1975.
[2] Z. W. Birnbaum and O. M. Klose. Bounds for the Variance of the Mann-Whitney
Statistic. Annals of Mathematical Statistics, 38, 1957.
[3] J-H. Chauchat, R. Rakotomalala, M. Carloz, and C. Pelletier. Targeting Customer
Groups using Gain and Cost Matrix; a Marketing Application. Technical report, ERIC
Laboratory - University of Lyon 2, 2001.
[4] Corinna Cortes and Mehryar Mohri. AUC Optimization vs. Error Rate Minimization.
In Advances in Neural Information Processing Systems (NIPS 2003), volume 16, Vancouver, Canada, 2004. MIT Press.
[5] D. Van Dantzig. On the Consistency and Power of Wilcoxon?s Two Sample Test. In
Koninklijke Nederlandse Akademie van Weterschappen, Series A, volume 54, 1915.
[6] J. P. Egan. Signal Detection Theory and ROC Analysis. Academic Press, 1975.
[7] C. Ferri, P. Flach, and J. Hern?andez-Orallo. Learning Decision Trees Using the Area
Under the ROC Curve. In Proceedings of the 19th International Conference on Machine Learning. Morgan Kaufmann, 2002.
[8] Yoav Freund and Robert E. Schapire. A Decision Theoretical Generalization of OnLine Learning and an Application to Boosting. In Proceedings of the Second European Conference on Computational Learning Theory, volume 2, 1995.
[9] Ronald L. Graham, Donald E. Knuth, and Oren Patashnik. Concrete Mathematics.
Addison-Wesley, Reading, Massachusetts, 1994.
[10] J. A. Hanley and B. J. McNeil. The Meaning and Use of the Area under a Receiver
Operating Characteristic (ROC) Curve. Radiology, 1982.
[11] E. L. Lehmann. Nonparametrics: Statistical Methods Based on Ranks. Holden-Day,
San Francisco, California, 1975.
[12] M. C. Mozer, R. Dodier, M. D. Colagrosso, C. Guerra-Salcedo, and R. Wolniewicz.
Prodding the ROC Curve: Constrained Optimization of Classifier Performance. In
Neural Information Processing Systems (NIPS 2002). MIT Press, 2002.
[13] C. Perlich, F. Provost, and J. Simonoff. Tree Induction vs. Logistic Regression: A
Learning Curve Analysis. Journal of Machine Learning Research, 2003.
[14] F. Provost and T. Fawcett. Analysis and Visualization of Classifier Performance:
Comparison under Imprecise Class and Cost Distribution. In Proceedings of the Third
International Conference on Knowledge Discovery and Data Mining. AAAI, 1997.
[15] Saharon Rosset. Ranking-Methods for Flexible Evaluation and Efficient Comparison
of 2-Class Models. Master?s thesis, Tel-Aviv University, 1999.
[16] L. Yan, R. Dodier, M. C. Mozer, and R. Wolniewicz. Optimizing Classifier Performance via the Wilcoxon-Mann-Whitney Statistics. In Proceedings of the International Conference on Machine Learning, 2003.
| 2645 |@word repository:2 version:1 briefly:2 polynomial:1 flach:1 km:3 crucially:1 salcedo:1 q1:2 thereby:1 fortuitous:1 moment:1 series:1 score:5 document:2 existing:3 com:1 z2:1 comparing:1 assigning:1 dx:1 must:1 readily:1 ronald:1 plot:2 v:2 selected:2 classier:1 provides:4 math:1 boosting:1 mathematical:1 ik:11 introduce:1 x0:7 pairwise:1 indeed:1 expected:9 inspired:1 relying:1 actual:1 lyon:1 considering:1 provided:1 underlying:1 questionable:2 stricter:1 exactly:2 classifier:16 k2:5 medical:1 yn:1 positive:24 limit:1 despite:1 ak:12 establishing:1 dantzig:1 suggests:1 limited:1 statistically:1 practical:4 acknowledgment:1 testing:1 yj:1 practice:1 block:1 x3:1 area:8 empirical:1 akademie:1 yan:1 significantly:1 imprecise:1 confidence:51 donald:1 cannot:1 close:1 selection:1 targeting:1 mea:1 py:5 optimize:1 equivalent:1 demonstrated:1 center:1 customer:1 starting:1 simplicity:1 m2:8 rule:1 bamber:1 annals:1 exact:4 origin:1 dodier:2 bottom:2 indep:3 ordering:1 decrease:1 principled:2 mozer:2 depend:2 tight:4 eric:1 k0:6 various:4 widely:1 favor:1 statistic:8 radiology:1 online:1 product:1 relevant:1 produce:1 perfect:1 help:1 derive:1 depending:1 measured:1 dep:3 eq:5 p2:3 strong:2 c:1 closely:1 mann:6 require:1 fix:1 andez:1 generalization:1 preliminary:1 proposition:9 probable:1 tighter:6 strictly:1 hold:3 considered:1 credit:1 normal:2 exp:1 k3:1 predict:1 pointing:1 m0:2 patashnik:1 a2:3 combinatorial:5 radio:1 kq0:1 equiprobability:1 weighted:1 minimization:1 mit:2 always:1 pn:1 corollary:4 derived:7 properly:2 improvement:1 rank:5 consistently:1 ferri:1 contrast:2 perlich:1 sense:1 dependent:11 typically:1 holden:1 a0:2 classification:23 flexible:1 denoted:1 constrained:1 uc:2 equal:1 represents:1 constitutes:1 report:4 simplify:1 few:1 equiprobable:2 randomly:5 maintain:1 detection:2 interest:2 highly:2 mining:2 evaluation:2 tree:2 loosely:1 penalizes:1 egan:1 theoretical:2 increased:1 whitney:7 yoav:1 cost:2 deviation:12 too:2 reported:1 rosset:2 international:3 connecting:1 concrete:1 thesis:1 aaai:1 nm:4 admit:1 inefficient:1 leading:2 colagrosso:1 coefficient:1 z12:1 ranking:15 depends:3 ad:1 closed:1 bayes:1 accuracy:2 variance:29 characteristic:3 efficiently:1 kaufmann:1 yield:2 weak:1 produced:1 finer:1 straight:2 definition:4 e2:5 proof:7 associated:2 irvine:2 gain:1 massachusetts:1 knowledge:1 organized:1 wesley:1 originally:1 courant:1 higher:5 day:1 adaboost:1 nonparametrics:1 just:1 marketing:1 google:2 logistic:1 quality:5 yeast:1 aviv:1 name:1 validity:2 true:1 q0:1 laboratory:1 auc:60 criterion:2 demonstrate:1 saharon:2 meaning:1 novel:1 empirically:1 overview:1 volume:3 extend:1 discussed:2 orallo:1 significant:1 consistency:1 pm:1 z4:1 mathematics:1 longer:1 operating:3 wilcoxon:6 optimizing:1 certain:1 inequality:4 binary:7 captured:1 morgan:1 determine:8 signal:3 desirable:1 technical:1 adapt:1 calculation:2 academic:1 equally:1 e1:5 a1:1 wolniewicz:2 regression:1 ae:2 expectation:6 fawcett:1 oren:1 justified:1 interval:51 crucial:2 contrarily:1 sure:1 integer:1 gave:2 zi:1 misclassifications:2 psychology:1 idea:1 chebyshev:3 expression:17 returned:1 york:2 useful:3 reduced:3 schapire:2 estimated:1 dominance:1 group:1 demonstrating:2 threshold:1 birnbaum:1 verified:1 graph:2 mcneil:2 year:1 sum:1 lehmann:1 master:1 decision:3 graham:1 bound:2 internet:1 pxy:1 x2:4 aspect:1 min:4 extremely:1 px:5 relatively:3 pelletier:1 describes:2 rob:1 making:1 equation:1 visualization:1 hern:1 loose:3 ordinal:1 addison:1 available:2 alternative:1 corinna:3 customary:1 binomial:6 top:2 exploit:1 giving:2 hanley:7 k1:2 practicality:1 already:2 mx:1 thank:1 topic:1 argue:3 induction:1 assuming:1 z3:1 ratio:1 equivalently:1 difficult:1 unfortunately:1 robert:1 pima:1 pregibon:1 claiming:1 broadway:2 negative:23 design:1 upper:1 datasets:2 y1:1 provost:2 sweeping:1 community:1 canada:1 introduced:1 extensive:1 connection:2 z1:1 engine:1 california:1 narrow:1 unsafe:1 nip:2 below:1 xm:1 guerra:1 reading:1 including:1 max:7 belief:1 ia:7 critical:1 daryl:1 ranked:1 natural:1 power:1 indicator:1 mn:10 risky:1 xq:5 review:1 literature:5 discovery:1 checking:1 vancouver:1 determining:1 law:4 freund:1 fully:1 interesting:2 degree:2 course:2 mohri:3 last:2 prodding:1 bias:1 institute:1 fall:1 benefit:2 van:2 curve:14 avoids:1 author:1 commonly:1 made:1 san:1 simplified:1 simonoff:1 reproduces:1 receiver:3 francisco:1 xi:1 search:1 continuous:1 table:3 tel:1 mehryar:2 complex:1 european:1 significance:2 pk:6 main:1 motivation:1 n2:7 x1:3 kn0:2 fig:2 roc:12 ny:2 a21:1 exponential:2 third:1 formula:1 theorem:6 specific:2 nyu:2 list:2 cortes:2 ionosphere:1 essential:2 false:3 importance:1 knuth:1 nk:3 likely:1 ordered:1 partially:1 viewed:3 identity:6 determined:1 uneven:1 violated:1 incorporate:1 |
1,812 | 2,646 | Face Detection ? Efficient and Rank Deficient
Wolf Kienzle, G?okhan Bak?r, Matthias Franz and Bernhard Scho? lkopf
Max-Planck-Institute for Biological Cybernetics
Spemannstr. 38, D-72076 T?ubingen, Germany
{kienzle, gb, mof, bs}@tuebingen.mpg.de
Abstract
This paper proposes a method for computing fast approximations to support vector decision functions in the field of object detection. In the
present approach we are building on an existing algorithm where the set
of support vectors is replaced by a smaller, so-called reduced set of synthesized input space points. In contrast to the existing method that finds
the reduced set via unconstrained optimization, we impose a structural
constraint on the synthetic points such that the resulting approximations
can be evaluated via separable filters. For applications that require scanning large images, this decreases the computational complexity by a significant amount. Experimental results show that in face detection, rank
deficient approximations are 4 to 6 times faster than unconstrained reduced set systems.
1
Introduction
It has been shown that support vector machines (SVMs) provide state-of-the-art accuracies
in object detection. In time-critical applications, however, they are of limited use due to
their computationally expensive decision functions. In particular, the time complexity of
an SVM classification operation is characterized by two parameters. First, it is linear in the
number of support vectors (SVs). Second, it scales with the number of operations needed
for computing the similarity between an SV and the input, i.e. the complexity of the kernel
function. When classifying image patches of size h ? w using plain gray value features, the
decision function requires an h ? w dimensional dot product for each SV. As the patch size
increases, these computations become extremely expensive. As an example, the evaluation
of a single 20 ? 20 patch on a 320 ? 240 image at 25 frames per second already requires
660 million operations per second.
In the past, research towards speeding up kernel expansions has focused exclusively on the
first issue, i.e. on how to reduce the number of expansion points (SVs) [1, 2]. In [2], Burges
introduced a method that, for a given SVM, creates a set of so-called reduced set vectors
(RSVs) that approximate the decision function. This approach has been successfully applied in the image classification domain ? speedups on the order of 10 to 30 have been reported [2, 3, 4] while the full accuracy was retained. Additionally, for strongly unbalanced
classification problems such as face detection, the average number of RSV evaluations can
be further reduced using cascaded classifiers [5, 6, 7]. Unfortunately, the above example
illustrates that even with as few as three RSVs on average (as in [5]), such systems are not
competitive for time-critical applications.
The present work focuses on the second issue, i.e. the high computational cost of the kernel
evaluations. While this could be remedied by switching to a sparser image representation
(e.g. a wavelet basis), one could argue that in connection with SVMs, not only are plain
gray values straightforward to use, but they have shown to outperform Haar wavelets and
gradients in the face detection domain [8]. Alternatively, in [9], the authors suggest to compute the costly correlations in the frequency domain. In this paper, we develop a method
that combines the simplicity of gray value correlations with the speed advantage of more
sophisticated image representations. To this end, we borrow an idea from image processing:
by constraining the RSVs to have a special structure, they can be evaluated via separable
convolutions. This works for most standard kernels (e.g. linear, polynomial, Gaussian and
sigmoid) and decreases the average computational complexity of the RSV evaluations from
O(h ? w) to O(r ? (h + w)), where r is a small number that allows the user to balance between speed and accuracy. To evaluate our approach, we examine the performance of these
approximations on the MIT+CMU face detection database (used in [10, 8, 5, 6]).
2
Burges? method for reduced set approximations
The present section briefly describes Burges? reduced set method [2] on which our work is
based. For reasons that will become clear below, h ? w image patches are written as h ? w
matrices (denoted by bold capital letters) whose entries are the respective pixel intensities.
In this paper, we refer to this as the image-matrix notation.
Assume that an SVM has been successfully trained on the problem at hand. Let
{X1 , . . . Xm } denote the set of SVs, {?1 , . . . ?m } the corresponding coefficients, k(?, ?)
the kernel function and b the bias of the SVM solution. The decision rule for a test pattern
X reads
!
m
X
f (X) = sgn
yi ?i k(Xi , X) + b .
(1)
i=1
In SVMs, the decision surface induced by f corresponds to a hyperplane in the reproducing
kernel Hilbert space (RKHS) associated with k. The corresponding normal
?=
m
X
yi ?i k(Xi , ?)
(2)
i=1
can be approximated using a smaller, so-called reduced set (RS) {Z1 , . . . Zm0 } of size
m0 < m, i.e. an approximation to ? of the form
0
0
? =
m
X
?i k(Zi , ?).
(3)
i=1
This speeds up the decision process by a factor of m/m0 . To find such ?0 , we fix a desired
set size m0 and solve
min k? ? ?0 k2RKHS
(4)
for ?i and Zi . Here, k ? kRKHS denotes the Euclidian norm in the RKHS. The resulting RS
decision function f 0 is then given by
? 0
?
m
X
f 0 (X) = sgn ?
?i k(Zi , X) + b? .
(5)
i=1
In practice, ?i , Zi are found using a gradient based optimization technique. Details can be
found in [2].
3
From separable filters to rank deficient reduced sets
We now describe the concept of separable filters in image processing and show how this
idea extends to a broader class of linear filters and to a special class of nonlinear filters,
namely those used by SVM decision functions. Using the image-matrix notation, it will
become clear that the separability property boils down to a matrix rank constraint.
3.1
Linear separable filters
Applying a linear filter to an image amounts to a two-dimensional convolution of the image
with the impulse response of the filter. In particular, if I is the input image, H the impulse
response, i.e. the filter mask, and J the output image, then
J = I ? H.
(6)
If H has size h ? w, the convolution requires O(h ? w) operations for each output pixel.
However, in special cases where H can be decomposed into two column vectors a and b,
such that
H = ab>
(7)
holds, we can rewrite (6) as
J = [I ? a] ? b> ,
(8)
>
>
since the convolution is associative and in this case, ab = a ? b . This splits the original
problem (6) into two convolution operations with masks of size h?1 and 1?w, respectively.
As a result, if a linear filter is separable in the sense of equation (7), the computational
complexity of the filtering operation can be reduced from O(h ? w) to O(h + w) per pixel
by computing (8) instead of (6).
3.2
Linear rank deficient filters
In view of (7) being equivalent to rank(H) ? 1, we now generalize the above concept to
linear filters with low rank impulse responses. Consider the singular value decomposition
(SVD) of the h ? w matrix H,
H = USV> ,
(9)
and recall that U and V are orthogonal matrices of size h ? h and w ? w, respectively,
whereas S is diagonal (the diagonal entries are the singular values) and has size h ? w.
Now let r = rank(H). Due to rank(S) = rank(H), we may write H as a sum of r rank
one matrices
r
X
s i u i vi >
(10)
H=
i=1
where si denotes the ith singular value of H and ui , vi are the iths columns of U and V
(i.e. the ith singular vectors of the matrix H), respectively. As a result, the corresponding linear filter can be evaluated (analogously to (8)) as the weighted sum of r separable
convolutions
r
X
J=
si [I ? ui ] ? vi >
(11)
i=1
and the computational complexity drops from O(h ? w) to O(r ? (h + w)) per output
pixel. Not surprisingly, the speed benefit depends on r, which can be seen to measure the
structural complexity1 of H. For square matrices (w = h) for instance, (11) does not give
any speedup compared to (6) if r > w/2.
1
In other words, the flatter the spectrum of HH> , the less benefit can be expected from (11).
3.3
Nonlinear rank deficient filters and reduced sets
Due to the fact that in 2D, correlation is identical with convolution if the filter mask is
rotated by 180 degrees (and vice versa), we can apply the above idea to any image filter
f (X) = g(c(H, X)) where g is an arbitrary nonlinear function and c(H, X) denotes the
correlation between images patches X and H (both of size h ? w). In SVMs this amounts
to using a kernel of the form
k(H, X) = g(c(H, X)).
(12)
If H has rank r, we may split the kernel evaluation into r separable correlations plus a
scalar nonlinearity. As a result, if the RSVs in a kernel expansion such as (5) satisfy
this constraint, the average computational complexity decreases from O(m0 ? h ? w) to
O(m0 ? r ? (h + w)) per output pixel. This concept works for many off-the-shelf kernels
used in SVMs. While linear, polynomial and sigmoid kernels are defined as functions of
input space dot products and therefore immediately satisfy equation (12), the idea applies
to kernels based on the Euclidian distance as well. For instance, the Gaussian kernel reads
k(H, X) = exp(?(c(X, X) ? 2c(H, X) + c(H, H))).
(13)
Here, the middle term is the correlation which we are going to evaluate via separable filters.
The first term is independent of the SVs ? it can be efficiently pre-computed and stored in
a separate image. The last term is merely a constant scalar independent of the image data.
Finally, note that these kernels are usually defined on vectors. Nevertheless, we can use
our image-matrix notation due to the fact that the squared Euclidian distance between two
vectors of gray values x and z may be written as
kx ? zk2 = kX ? Zk2F ,
(14)
whereas the dot product amounts to
1
x> z =
kXk2F + kZk2F ? kX ? Zk2F ,
(15)
2
where X and Z are the corresponding image patches and k ? kF is the Frobenius norm for
matrices.
4
Finding rank deficient reduced sets
In our approach we consider a special class of the approximations given by (3), namely
those where the RSVs can be evaluated efficiently via separable correlations. In order to
obtain such approximations, we use a constrained version of Burges? method. In particular,
we restrict the RSV search space to the manifold spanned by all image patches that ?
viewed as matrices ? have a fixed, small rank r (which is to be chosen a priori by the user).
To this end, the Zi in equation (3) are replaced by their singular value decompositions
Z i ? U i S i Vi > .
(16)
The rank constraint can then be imposed by allowing only the first r diagonal elements of
Si to be non-zero. Note that this boils down to using an approximation of the form
0
?0r =
m
X
?i k(Ui,r Si,r Vi,r > , ?)
(17)
i=1
with Si,r being r ? r (diagonal) and Ui,r , Vi,r being h ? r, w ? r (orthogonal2 ) matrices,
respectively. Analogously to (4) we fix m0 and r and find Si,r , Ui,r , Vi,r and ?i that minimize the approximation error ? = k? ? ?0r k2RKHS . The minimization problem is solved via
2
In this paper we call a non-square matrix orthogonal if its columns are pairwise orthogonal and
have unit length.
gradient decent. Note that when computing gradients, the image-matrix notation (together
with (14) or (15), and the equality kXk2F = tr(XX> )) allows a straightforward computation of the kernel derivatives w.r.t. the components of the decomposed RSV image patches,
i.e. the row, column and scale information in Vi,r , Ui,r and Si,r , respectively. However,
while the update rules for ?i and Si,r follow immediately from the respective derivatives,
care must be taken in order to keep Ui,r and Vi,r orthogonal during optimization. This
can be achieved through re-orthogonalization of these matrices after each gradient step.
In our current implementation, however, we perform those updates subject to the so-called
Stiefel constraints [11]. Intuitively, this amounts to rotating (rather than translating) the
columns of Ui,r and Vi,r , which ensures that the resulting matrices are still orthogonal,
i.e. lie on the Stiefel manifold. Let S(h, r) be the manifold of orthogonal h ? r matrices,
the (h, r)-Stiefel manifold. Further, let U?
i,r denote an orthogonal basis for the orthogonal complement of the subspace spanned by the columns of Ui,r . Now, given the ?free?
gradient G = ??/?Ui,r we compute the ?constrained? gradient
? = G ? Ui,r G> Ui,r ,
G
(18)
which is the projection of G onto the tangent space of S(h, r) at Ui,r . The desired rotation
is then given [11] by the (matrix) exponential of the h ? h skew-symmetric matrix
!
? > Ui,r ?(G
? > U? )>
G
i,r
A=t?
(19)
? > U?
G
0
i,r
where t is a user-defined step size parameter. For details, see [11]. A Matlab library is
available at [12].
5
Experiments
This section shows the results of two experiments. The first part illustrates the behavior of
rank deficient approximations for a face detection SVM in terms of the convergence rate
and classification accuracy for different values of r. In the second part, we show how an
actual face detection system, similar to that presented in [5], can be sped up using rank
deficient RSVs. In both experiments we used the same training and validation set. It
consisted of 19 ? 19 gray level image patches containing 16081 manually collected faces
(3194 of them kindly provided by Sami Romdhani) and 42972 non-faces automatically
collected from a set of 206 background scenes. Each patch was normalized to zero mean
and unit variance. The set was split into a training set (13331 faces and 35827 non-faces)
and a validation set (2687 faces and 7145 non-faces). We trained a 1-norm soft margin
SVM on the training set using a Gaussian kernel with ? = 10. The regularization constant
C was set to 1. The resulting decision function (1) achieved a hit rate of 97.3% at 1.0%
false positives on the validation set using m = 6910 SVs. This solution served as the
approximation target ? (see equation (2)) during the experiments described below.
5.1
Rank deficient faces
In order to see how m0 and r affect the accuracy of our approximations, we compute rank
deficient reduced sets for m0 = 1 . . . 32 and r = 1 . . . 3 (the left array in Figure 1 illustrates
the actual appearance of rank deficient RSVs for the m0 = 6 case). Accuracy of the resulting decision functions is measured in ROC score (the area under the ROC curve) on the
validation set. For the full SVM, this amounts to 0.99. The results for our approximations
are depicted in Figure 2. As expected, we need a larger number of rank deficient RSVs
than unconstrained RSVs to obtain similar classification accuracies, especially for small
r. Nevertheless, the experiment points out two advantages of our method. First, a rank as
m' = 1
r=ful
=
+
+
r=3
=
+
+
r=2
=
+
r=1
r=1
r=2
r=3
r=full
m' = 6
=
+ ...
Figure 1: Rank deficient faces. The left array shows the RSVs (Zi ) of the unconstrained
(top row) and constrained (r decreases from 3 to 1 down the remaining rows) approximations for m0 = 6. Interestingly, the r = 3 RSVs are already able to capture face-like
structures. This supports the fact that the classification accuracy for r = 3 is similar to
that of the unconstrained approximations (cf. Figure 2, left plot). The right array shows the
m0 = 1 RSVs (r = full, 3, 2, 1, top to bottom row) and their decomposition into rank one
matrices according to (10). For the unconstrained RSV (first row) it shows an approximate
(truncated) expansion based on the three leading singular vectors. While for r = 3 the
decomposition is indeed similar to the truncated SVD, note how this similarity decreases
for r = 2, 1. This illustrates that the approach is clearly different from simply finding unconstrained RSVs and then imposing the rank constraint via SVD (in fact, the norm (4) is
smaller for the r = 1 RSV than for the leading singular vector of the r = full RSV).
low as three seems already sufficient for our face detection SVM in the sense that for equal
sizes m0 there is no significant loss in accuracy compared to the unconstrained approximation (at least for m0 > 2). The associated speed benefit over unconstrained RSVs is shown
in the right plot of Figure 2: the rank three approximations achieve accuracies similar to
the unconstrained functions, while the number of operations reduces to less than a third.
Second, while for unconstrained RSVs there is no solution with a number of operations
smaller than h ? w = 361 (in the right plot, this is the region beyond the left end of the solid
line), there exist rank deficient functions which are not only much faster than this, but yield
considerably higher accuracies. This property will be exploited in the next experiment.
5.2
A cascade-based face detection system
In this experiment we built a cascade-based face detection system similar to [5, 6], i.e.
a cascade of RSV approximations of increasing size m0 . As the benefit of a cascaded
classifier heavily depends on the speed of the first classifier which has to be evaluated on
the whole image [5, 6], our system uses a rank deficient approximation as the first stage.
Based on the previous experiment, we chose the m0 = 3, r = 1 classifier. Note that this
function yields an ROC score of 0.9 using 114 multiply-adds, whereas the simplest possible
unconstrained approximation m0 = 1, r = full needs 361 multiply-adds to achieve a ROC
score of only 0.83 (cf. Figure 2). In particular, if the threshold of the first stage is set to
yield a hit rate of 95% on the validation set, scanning the MIT+CMU set (130 images, 507
faces) with m0 = 3, r = 1 discards 91.5% of the false positives, whereas the m0 = 1,
r = full can only reject 70.2%. At the same time, when scanning a 320 ? 240 image 3 , the
three separable convolutions plus nonlinearity require 55ms, whereas the single, full kernel
evaluation takes 208ms on a Pentium 4 with 2.8 GHz. Moreover, for the unconstrained
3
For multi-scale processing the detectors are evaluated on an image pyramid with 12 different
scales using a scale decay of 0.75. This amounts to scanning 140158 patches for a 320 ? 240 image.
1
0.9
0.9
0.8
r=1
r=2
r=3
r=full
0.7
0.6
0
10
1
10
#RSVs (m')
ROC score
ROC score
1
0.8
r=1
r=2
r=3
r=full
0.7
0.6
2
10
3
10
4
10
#operations (m'?r?(h+w))
Figure 2: Effect of the rank parameter r on classification accuracies. The left plots shows
the ROC score of the rank deficient RSV approximations (cf. Section 4) for varying set sizes
(m0 = 1 . . . 32, on a logarithmic scale) and ranks (r = 1 . . . 3). Additionally, the solid line
shows the accuracy of the RSVs without rank constraint (cf. Section 2), here denoted by
r = full. The right plot shows the same four curves, but plotted against the number of
operations needed for the evaluation of the corresponding decision function when scanning
large images (i.e. m0 ? r ? (h + w) with h = w = 19), also on a logarithmic scale.
Figure 3: A sample output from our demonstration system (running at 14 frames per second).
In this implementation, we reduced the number
of false positives by adjusting the threshold of
the final classifier. Although this reduces the
number of detections as well, the results are still
satisfactory. This is probably due to the fact that
the MIT+CMU set contains several images of
very low quality that are not likely to occur in
our setting, using a good USB camera.
cascade to catch up in terms of accuracy, the (at least) m0 = 2, r = full classifier (also
with an ROC score of roughly 0.9) should be applied afterwards, requiring another 0.3 ?
2 ? 208 ms ? 125ms.
The subsequent stages of our system consist of unconstrained RSV approximations of size
m0 = 4, 8, 16, 32, respectively. These sizes were chosen such that the number of false
positives roughly halves after each stage, while the number of correct detections remains
close to 95% on the validation set (with the decision thresholds adjusted accordingly). To
eliminate redundant detections, we combine overlapping detections via averaging of position and size if they are closer than 0.15 times the estimated patch size. This system yields
93.1% correct detections and 0.034% false positives on the MIT+CMU set. The current
system was incorporated into a demo application (Figure 3). For optimal performance, we
re-compiled our system using the Intel compiler (ICC). The application now classifies a
320x240 image within 54ms (vs. 238ms with full rank RSVs only) on a 2.8 GHz PC. To
further reduce the number of false positives, additional bootstrapped (as in [5]) stages need
to be added to the cascade. Note that this will not significantly affect the speed of our system (currently 14 frames per second) since 0.034% false positives amounts to merely 47
patches to be processed by subsequent classifiers.
6
Discussion
We have presented a new reduced set method for SVMs in image processing, which creates sparse kernel expansions that can be evaluated via separable filters. To this end, the
user-defined rank (the number of separable filters into which the RSVs are decomposed)
provides a mechanism to control the tradeoff between accuracy and speed of the resulting
approximation. Our experiments show that for face detection, the use of rank deficient
RSVs leads to a significant speedup without losing accuracy. Especially when rough approximations are required, our method gives superior results compared to the existing reduced set methods since it allows for a finer granularity which is vital in cascade-based
detection systems. Another property of our approach is simplicity. At run-time, rank deficient RSVs can be used together with unconstrained RSVs or SVs using the same canonical
image representation. As a result, the required changes in existing code, such as in [5], are
small. In addition, our approach allows the use of off-the-shelf image processing libraries
for separable convolutions. Since such operations are essential in image processing, there
exist many (often highly optimized) implementations. Finally, the method can well be used
to train a neural network, i.e. to go directly from the training data to a sparse, separable
function as opposed to taking the SVM ?detour?. A comparison of that approach to the
present one, however, remains to be done.
References
[1] E. Osuna and F. Girosi. Reducing the run-time complexity in support vector machines. In
B. Sch?olkopf, C. J. C. Burges, and A. J. Smola, editors, Advances in Kernel Methods ? Support
Vector Learning, pages 271?284, Cambridge, MA, 1999. MIT Press.
[2] C. J. C. Burges. Simplified support vector decision rules. In International Conference on
Machine Learning, pages 71?77, 1996.
[3] C. J. C. Burges and B. Scho? lkopf. Improving the accuracy and speed of support vector machines. In M. C. Mozer, M. I. Jordan, and T. Petsche, editors, Advances in Neural Information
Processing Systems, volume 9, page 375. MIT Press, 1997.
[4] E. Osuna, R. Freund, and F. Girosi. Training support vector machines: an application to face
detection. In Proceedings IEEE Conference on Computer Vision and Pattern Recognition, 1997.
[5] S. Romdhani, P. Torr, B.Sch?olkopf, and A. Blake. Computationally efficient face detection. In
Proceedings of the International Conference on Computer Vision, pages 695?700, 2001.
[6] P. Viola and M. Jones. Rapid object detection using a boosted cascade of simple features. In
Proceedings IEEE Conference on Computer Vision and Pattern Recognition, 2001.
[7] G. Blanchard and D. Geman. Hierarchical testing designs for pattern recognition. Technical
Report 2003-07, Universit Paris-Sud, 2003.
[8] B. Heisele, T. Poggio, and M. Pontil. Face detection in still gray images. AI Memo 1687, MIT,
May 2000. CBCL Memo 187.
[9] S. Ben-Yacoub, B. Fasel, and J. Luettin. Fast face detection using MLP and FFT. In Proceedings
International Conference on Audio and Video-based Biometric Person Authentication, 1999.
[10] H. A. Rowley, S. Baluja, and T. Kanade. Neural network-based face detection. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(1):23?38, January 1998.
[11] A. Edelman, T. Arias, and S. Smith. The geometry of algorithms with orthogonality constraints.
SIAM Journal on Matrix Analysis and Applications, 20:303?353, 1998.
[12] RDRSLIB ? a matlab library for rank deficient reduced sets in object detection,
http://www.kyb.mpg.de/bs/people/kienzle/rdrs/rdrs.htm.
| 2646 |@word middle:1 version:1 briefly:1 polynomial:2 norm:4 seems:1 r:2 decomposition:4 euclidian:3 tr:1 solid:2 contains:1 exclusively:1 score:7 rkhs:2 interestingly:1 bootstrapped:1 past:1 existing:4 current:2 si:8 written:2 must:1 subsequent:2 girosi:2 kyb:1 x240:1 drop:1 plot:5 update:2 v:1 half:1 intelligence:1 accordingly:1 ith:2 smith:1 provides:1 become:3 edelman:1 combine:2 pairwise:1 mask:3 indeed:1 expected:2 roughly:2 mpg:2 examine:1 rapid:1 multi:1 sud:1 behavior:1 decomposed:3 automatically:1 actual:2 increasing:1 provided:1 xx:1 notation:4 moreover:1 classifies:1 finding:2 ful:1 universit:1 classifier:7 hit:2 control:1 unit:2 planck:1 positive:7 fasel:1 switching:1 plus:2 chose:1 kxk2f:2 limited:1 camera:1 testing:1 practice:1 heisele:1 pontil:1 area:1 cascade:7 significantly:1 reject:1 projection:1 word:1 pre:1 suggest:1 onto:1 close:1 applying:1 www:1 equivalent:1 imposed:1 straightforward:2 go:1 focused:1 simplicity:2 immediately:2 rule:3 array:3 borrow:1 spanned:2 target:1 heavily:1 user:4 losing:1 us:1 element:1 expensive:2 approximated:1 recognition:3 geman:1 database:1 bottom:1 solved:1 capture:1 region:1 ensures:1 decrease:5 mozer:1 complexity:8 ui:14 rowley:1 trained:2 rewrite:1 creates:2 basis:2 htm:1 train:1 fast:2 describe:1 whose:1 larger:1 solve:1 final:1 associative:1 advantage:2 matthias:1 product:3 achieve:2 frobenius:1 olkopf:2 convergence:1 rotated:1 object:4 ben:1 develop:1 measured:1 bak:1 correct:2 filter:19 sgn:2 translating:1 require:2 fix:2 biological:1 adjusted:1 hold:1 blake:1 normal:1 exp:1 cbcl:1 m0:22 currently:1 vice:1 successfully:2 weighted:1 minimization:1 mit:7 clearly:1 rough:1 gaussian:3 rather:1 shelf:2 boosted:1 varying:1 broader:1 focus:1 rank:38 contrast:1 pentium:1 sense:2 eliminate:1 going:1 germany:1 pixel:5 issue:2 classification:7 biometric:1 denoted:2 priori:1 proposes:1 art:1 special:4 constrained:3 field:1 equal:1 manually:1 identical:1 jones:1 report:1 few:1 replaced:2 geometry:1 ab:2 detection:26 mlp:1 highly:1 multiply:2 evaluation:7 pc:1 closer:1 poggio:1 respective:2 orthogonal:8 detour:1 rotating:1 desired:2 re:2 plotted:1 instance:2 column:6 soft:1 cost:1 entry:2 reported:1 stored:1 scanning:5 sv:2 synthetic:1 considerably:1 person:1 international:3 siam:1 off:2 analogously:2 together:2 squared:1 containing:1 opposed:1 derivative:2 leading:2 de:2 bold:1 flatter:1 coefficient:1 blanchard:1 satisfy:2 vi:10 depends:2 view:1 compiler:1 competitive:1 minimize:1 square:2 accuracy:17 variance:1 efficiently:2 yield:4 generalize:1 lkopf:2 served:1 usb:1 cybernetics:1 finer:1 detector:1 romdhani:2 against:1 frequency:1 associated:2 boil:2 adjusting:1 recall:1 hilbert:1 sophisticated:1 higher:1 follow:1 response:3 evaluated:7 done:1 strongly:1 stage:5 smola:1 correlation:7 hand:1 nonlinear:3 overlapping:1 quality:1 gray:6 impulse:3 building:1 effect:1 concept:3 consisted:1 normalized:1 requiring:1 equality:1 regularization:1 read:2 symmetric:1 satisfactory:1 during:2 authentication:1 m:6 usv:1 orthogonalization:1 stiefel:3 image:38 scho:2 sigmoid:2 rotation:1 superior:1 sped:1 volume:1 million:1 synthesized:1 significant:3 refer:1 versa:1 imposing:1 cambridge:1 ai:1 unconstrained:15 nonlinearity:2 dot:3 similarity:2 surface:1 compiled:1 add:2 discard:1 ubingen:1 yi:2 exploited:1 seen:1 additional:1 care:1 impose:1 redundant:1 afterwards:1 full:13 reduces:2 technical:1 faster:2 characterized:1 okhan:1 kzk2f:1 vision:3 cmu:4 kernel:19 pyramid:1 achieved:2 luettin:1 whereas:5 background:1 addition:1 singular:7 sch:2 probably:1 induced:1 subject:1 deficient:19 spemannstr:1 jordan:1 call:1 structural:2 granularity:1 constraining:1 split:3 sami:1 decent:1 vital:1 fft:1 affect:2 zi:6 restrict:1 reduce:2 idea:4 tradeoff:1 gb:1 complexity1:1 svs:6 matlab:2 clear:2 amount:8 svms:6 processed:1 simplest:1 reduced:17 http:1 outperform:1 exist:2 canonical:1 estimated:1 per:7 write:1 four:1 nevertheless:2 threshold:3 capital:1 merely:2 sum:2 run:2 letter:1 extends:1 patch:13 decision:14 occur:1 constraint:8 orthogonality:1 scene:1 speed:9 extremely:1 min:1 separable:15 speedup:3 according:1 smaller:4 describes:1 separability:1 osuna:2 b:2 intuitively:1 taken:1 computationally:2 equation:4 remains:2 skew:1 mechanism:1 hh:1 needed:2 end:4 zk2:1 available:1 operation:11 apply:1 hierarchical:1 petsche:1 original:1 denotes:3 top:2 remaining:1 cf:4 running:1 especially:2 already:3 added:1 costly:1 diagonal:4 gradient:7 subspace:1 distance:2 separate:1 remedied:1 manifold:4 argue:1 collected:2 tuebingen:1 reason:1 length:1 code:1 retained:1 balance:1 demonstration:1 unfortunately:1 memo:2 implementation:3 design:1 zk2f:2 allowing:1 perform:1 convolution:9 truncated:2 january:1 viola:1 incorporated:1 frame:3 reproducing:1 arbitrary:1 intensity:1 introduced:1 complement:1 namely:2 required:2 paris:1 connection:1 z1:1 optimized:1 able:1 beyond:1 below:2 pattern:5 xm:1 usually:1 built:1 max:1 video:1 critical:2 haar:1 cascaded:2 library:3 catch:1 speeding:1 mof:1 tangent:1 kf:1 icc:1 freund:1 loss:1 zm0:1 filtering:1 validation:6 degree:1 sufficient:1 editor:2 classifying:1 row:5 surprisingly:1 last:1 free:1 bias:1 burges:7 institute:1 face:26 taking:1 sparse:2 benefit:4 ghz:2 curve:2 plain:2 author:1 franz:1 simplified:1 transaction:1 approximate:2 bernhard:1 rsv:10 keep:1 xi:2 demo:1 alternatively:1 spectrum:1 search:1 additionally:2 kanade:1 improving:1 expansion:5 domain:3 kindly:1 whole:1 x1:1 intel:1 roc:8 position:1 exponential:1 lie:1 third:1 wavelet:2 down:3 decay:1 svm:10 consist:1 essential:1 false:7 aria:1 illustrates:4 kx:3 margin:1 sparser:1 depicted:1 logarithmic:2 simply:1 appearance:1 likely:1 yacoub:1 scalar:2 applies:1 wolf:1 corresponds:1 ma:1 viewed:1 towards:1 change:1 baluja:1 torr:1 reducing:1 hyperplane:1 averaging:1 kienzle:3 called:4 experimental:1 svd:3 support:10 people:1 unbalanced:1 evaluate:2 audio:1 |
1,813 | 2,647 | Non-Local Manifold Tangent Learning
Yoshua Bengio and Martin Monperrus
Dept. IRO, Universit?e de Montr?eal
P.O. Box 6128, Downtown Branch, Montreal, H3C 3J7, Qc, Canada
{bengioy,monperrm}@iro.umontreal.ca
Abstract
We claim and present arguments to the effect that a large class of manifold learning algorithms that are essentially local and can be framed as
kernel learning algorithms will suffer from the curse of dimensionality,
at the dimension of the true underlying manifold. This observation suggests to explore non-local manifold learning algorithms which attempt to
discover shared structure in the tangent planes at different positions. A
criterion for such an algorithm is proposed and experiments estimating a
tangent plane prediction function are presented, showing its advantages
with respect to local manifold learning algorithms: it is able to generalize very far from training data (on learning handwritten character image
rotations), where a local non-parametric method fails.
1
Introduction
A central issue of generalization is how information from the training examples can be
used to make predictions about new examples, and without strong prior assumptions, i.e.
in non-parametric models, this may be fundamentally difficult as illustrated by the curse
of dimensionality. There has been in recent years a lot of work on unsupervised learning based on characterizing a possibly non-linear manifold near which the data would lie,
such as Locally Linear Embedding (LLE) (Roweis and Saul, 2000), Isomap (Tenenbaum,
de Silva and Langford, 2000), kernel Principal Components Analysis (PCA) (Sch o? lkopf,
Smola and M?uller, 1998), Laplacian Eigenmaps (Belkin and Niyogi, 2003), and Manifold Charting (Brand, 2003). These are all essentially non-parametric methods that can be
shown to be kernel methods with an adaptive kernel (Bengio et al., 2004), and which represent the manifold on the basis of local neighborhood relations, very often constructed using the nearest neighbors graph (the graph with one vertex per observed example, and arcs
between near neighbors). The above methods characterize the manifold through an embedding which associates each training example (an input object) with a low-dimensional
coordinate vector (the coordinates on the manifold). Other closely related methods characterize the manifold as well as ?noise? around it. Most of these methods consider the
density as a mixture of flattened Gaussians, e.g. mixtures of factor analyzers (Ghahramani
and Hinton, 1996), Manifold Parzen windows (Vincent and Bengio, 2003), and other local
PCA models such as mixtures of probabilistic PCA (Tipping and Bishop, 1999). This is
not an exhaustive list, and recent work also combines modeling through a mixture density
and dimensionality reduction (Teh and Roweis, 2003; Brand, 2003).
In this paper we claim that there is a fundamental weakness with such kernel methods, due
to the locality of learning: we show that the local tangent plane of the manifold at a point
x is defined based mostly on the near neighbors of x according to some possibly data-
dependent kernel KD . As a consequence, it is difficult with such methods to generalize to
new combinations of values x that are ?far? from the training examples xi , where being
?far? is a notion that should be understood in the context of several factors: the amount of
noise around the manifold (the examples do not lie exactly on the manifold), the curvature
of the manifold, and the dimensionality of the manifold. For example, if the manifold
curves quickly around x, neighbors need to be closer for a locally linear approximation to
be meaningful, which means that more data are needed. Dimensionality of the manifold
compounds that problem because the amount of data thus needed will grow exponentially
with it. Saying that y is ?far? from x means that y is not well represented by its projection
on the tangent plane at x.
In this paper we explore one way to address that problem, based on estimating the tangent
planes of the manifolds as a function of x, with parameters that can be estimated not only
from the data around x but from the whole dataset. Note that there can be more than one
manifold (e.g. in vision, one may imagine a different manifold for each ?class? of object),
but the structure of these manifolds may be related, something that many previous manifold
learning methods did not take advantage of. We present experiments on a variety of tasks
illustrating the weaknesses of the local manifold learning algorithms enumerated above.
The most striking result is that the model is able to generalize a notion of rotation learned
on one kind of image (digits) to a very different kind (alphabet characters), i.e. very far
from the training data.
2
Local Manifold Learning
By ?local manifold learning?, we mean a method that derives information about the local
structure of the manifold (i.e. implicitly its tangent directions) at x based mostly on the
training examples ?around? x. There is a large group of manifold learning methods (as
well as the spectral clustering methods) that share several characteristics, and can be seen
as data-dependent kernel PCA (Bengio et al., 2004). These include LLE (Roweis and Saul,
2000), Isomap (Tenenbaum, de Silva and Langford, 2000), kernel PCA (Sch o? lkopf, Smola
and M?uller, 1998) and Laplacian Eigenmaps (Belkin and Niyogi, 2003). They first build a
data-dependent Gram matrix M with n ? n entries KD (xi , xj ) where D = {x1 , . . . , xn }
is the training set and KD is a data-dependent kernel, and compute the eigenvectoreigenvalue pairs {(vk , ?k )} of M . The embedding of the training set is obtained directly
from the principal eigenvectors vk of M (the
q i-th element of vk gives the k-th coordinate of xi ?s embedding, possibly scaled by ?nk , i.e. ek (xi ) = vik ) and the embedding
for a new example
can be estimated using the Nystro? m formula (Bengio et al., 2004):
Pn
ek (x) = ?1k i=1 vki KD (x, xi ) for the k-th coordinate of x, where ?k is the k-th eigenq
value of M (the optional scaling by ?nk would also apply). The above equation says
that the embedding for a new example x is a local interpolation of the manifold coori)
dinates of its neighbors xi , with interpolating weights given by KD?(x,x
. To see more
k
clearly how the tangent plane may depend only on the neighbors of x, consider the relation between the tangent plane and the embedding function: the tangent plane at x is
k (x)
simply the subspace spanned by the vectors ?e?x
. In the case of very ?local? kernels
like that of LLE, spectral clustering with Gaussian kernel, Laplacian Eigenmaps or kernel
PCA with Gaussian kernel, that derivative only depends significantly on the near neighk (x)
bors of x. Consider for example kernel PCA with a Gaussian kernel: then ?e?x
can be
closely approximated by a linear combination of the difference vectors (x ? x j ) for xj
near x. The weights of that combination may depend on the whole data set, but if the
ambiant space has many more dimensions then the number of such ?near? neighbors of
x, this is a very strong locally determined constraint on the shape of the manifold. The
case of Isomap is less obvious but we show below that it is also local. Let D(a, b) denote
the graph geodesic distance going only through a, b and points from the training set. As
shown in (Bengio et al., 2004), thePcorresponding data-dependent kernel canPbe defined as
2
? i + D)
? where D
?i = 1
KD (x, xi ) = ? 21 (D(x, xi )2 ? n1 j D(x, xj )2 ? D
j D(xi , xj )
n
P
1
?=
?
and D
j Dj . Let N (x, xi ) denote the index j of the training set example xj that is
n
a neighbor of x that minimizes ||x ? xj || + D(xj , xi ). Then
?
?
X
X
(x ? xN (x,xj ) )
(x ? xN (x,xi ) )
?ek (x)
1
1
?
=
vki ?
D(x, xj )
? D(x, xi )
?x
?k i
n j
||x ? xN (x,xj ) ||
||x ? xN (x,xi ) ||
(1)
which is a linear combination of vectors (x ? xk ), where xk is a neighbor of x. This clearly
shows that the tangent plane at x associated with Isomap is also included in the subspace
spanned by the vectors (x ? xk ) where xk is a neighbor of x.
There are also a variety of local manifold learning algorithms which can be classified as
?mixtures of pancakes? (Ghahramani and Hinton, 1996; Tipping and Bishop, 1999; Vincent and Bengio, 2003; Teh and Roweis, 2003; Brand, 2003). These are generally mixtures
of Gaussians with a particular covariance structure. When the covariance matrix is approximated using its principal eigenvectors, this leads to ?local PCA? types of methods. For
these methods the local tangent directions directly correspond to the principal eigenvectors
of the local covariance matrices. Learning is also local since it is mostly the examples
around the Gaussian center that determine its covariance structure. The problem is not so
much due to the form of the density as a mixture of Gaussians. The problem is that the local
parameters (e.g. local principal directions) are estimated mostly based on local data. There
is usually a non-local interaction between the different Gaussians, but its role is mainly of
global coordination, e.g. where to set the Gaussian centers to allocate them properly where
there is data, and optionally how to orient the principal directions so as to obtain a globally
coherent coordinate system for embedding the data.
2.1
Where Local Manifold Learning Would Fail
It is easy to imagine at least four failure causes for local manifold learning methods, and
combining them will create even greater problems:
? Noise around the manifold: data are not exactly lying on the manifold. In the case
of non-linear manifolds, the presence of noise means that more data around each pancake
region will be needed to properly estimate the tangent directions of the manifold in that
region.
? High curvature of the manifold. Local manifold learning methods basically approximate the manifold by the union of many locally linear patches. For this to work, there must
be at least d close enough examples in each patch (more with noise). With a high curvature
manifold, more ? smaller ? patches will be needed, and the number of required patches
will grow exponentially with the dimensionality of the manifold. Consider for example the
manifold of translations of a high-contrast image.The tangent direction corresponds to the
change in image due a small translation, i.e. it is non-zero only at edges in the image. After
a one-pixel translation, the edges have moved by one pixel, and may not overlap much with
the edges of the original image if it had high contrast. This is indeed a very high curvature
manifold.
? High intrinsic dimension of the manifold. We have already seen that high manifold
dimensionality d is hurtful because O(d) examples are required in each patch and O(r d )
patches (for some r depending on curvature and noise) are necessary to span the manifold.
In the translation example, if the image resolution is increased, then many more training
images will be needed to capture the curvature around the translation manifold with locally
linear patches. Yet the physical phenomenon responsible for translation is expressed by a
simple equation, which does not get more complicated with increasing resolution.
? Presence of many manifolds with little data per manifold. In many real-world contexts there is not just one global manifold but a large number of manifolds which however
share something about their structure. A simple example is the manifold of transformations (view-point, position, lighting,...) of 3D objects in 2D images. There is one manifold
per object instance (corresponding to the successive application of small amounts of all
of these transformations). If there are only a few examples for each such class then it
is almost impossible to learn the manifold structures using only local manifold learning.
However, if the manifold structures are generated by a common underlying phenomenon
then a non-local manifold learning method could potentially learn all of these manifolds
and even generalize to manifolds for which a single instance is observed, as demonstrated
in the experiments below.
3
Non-Local Manifold Tangent Learning
Here we choose to characterize the manifolds in the data distribution through a matrixvalued function F (x) that predicts at x ? Rn a basis for the tangent plane of the manifold
near x, hence F (x) ? Rd?n for a d-dimensional manifold. Basically, F (x) specifies
?where? (in which directions) one expects to find near neighbors of x.
We are going to consider a simple supervised learning setting to train this function. As
with Isomap, we consider that the vectors (x ? xi ) with xi a near neighbor of x span a
noisy estimate of the manifold tangent space. We propose to use them to define a ?target?
for training F (x). In our experiments we simply collected the k nearest neighbors of each
example x, but better selection criteria might be devised. Points on the predicted tangent
subspace can be written F 0 (x)w with w ? Rd being local coordinates in the basis specified
by F (x). Several criteria are possible to match the neighbors differences with the subspace
defined by F (x). One that yields to simple analytic calculations is simply to minimize
the distance between the x ? xj vectors and their projection on the subspace defined by
F (x). The low-dimensional local coordinate vector wtj ? Rd that matches neighbor xj
of example xt is thus an extra free parameter that has to be optimized, but is obtained
analytically. The overall training criterion involves a double optimization over function F
and local coordinates wtj of what we call the relative projection error:
min
F,{wtj }
X X
t
j?N (xt )
||F 0 (xt )wtj ? (xt ? xj )||2
||xt ? xj ||2
(2)
where N (x) denotes the selected set of near neighbors of x. The normalization by ||x t ?
xj ||2 is to avoid giving more weight to the neighbors that are further away. The above
ratio amounts to minimizing the square of the sinus of the projection angle. To perform
the above minimization, we can do coordinate descent (which guarantees convergence to
a minimum), i.e. alternate changes in F and changes in w?s which at each step go down
the total criterion. Since the minimization over the w?s can be done separately for each
example xt and neighbor xj , it is equivalent to minimize
||F 0 (xt )wtj ? (xt ? xj )||2
||xt ? xj ||2
(3)
over vector wtj for each such pair (done analytically) and compute the gradient of the
above over F (or its parameters) to move F slightly (we used stochastic gradient on the
parameters of F ). The solution for wtj is obtained by solving the linear system
F (xt )F 0 (xt )wtj = F (xt )
(xt ? xj )
.
||xt ? xj ||2
(4)
In our implementation this is done robustly through a singular value decomposition
F 0 (xt ) = U SV 0 and wtj = B(xt ? xj ) where B can be precomputed for all the neighbors
Pd
of xt : B = ( k=1 1Sk > V.k V.k0 /Sk2 )F (xt ). The gradient of the criterion with respect to
the i-th row of F (xt ), holding wtj fixed, is simply
X
wtji
(F 0 (xt )w ? (xt ? xj ))
2
||x
?
x
||
t
j
j
(5)
where wtji is the i-th element of wtj . In practice, it is not necessary to store more than one
wtj vector at a time. In the experiments, F (?) is parameterized as a an ordinary one hidden
layer neural network with n inputs and d ? n outputs. It is trained by stochastic gradient
descent, one example xt at a time.
Although the above algorithm provides a characterization of the manifold, it does not directly provide an embedding nor a density function. However, once the tangent plane
function is trained, there are ways to use it to obtain all of the above. The simplest method
is to apply existing algorithms that provide both an embedding and a density function based
on a Gaussian mixture with pancake-like covariances. For example one could use (Teh and
Roweis, 2003) or (Brand, 2003), the local covariance matrix around x being constructed
from F 0 (x)diag(? 2 (x))F (x), where ?i2 (x) should estimate V ar(wi ) around x.
3.1
Previous Work on Non-Local Manifold Learning
The non-local manifold learning algorithm presented here (find F (?) which minimizes the
criterion in eq. 2) is similar to the one proposed in (Rao and Ruderman, 1999) to estimate the generator matrix of a Lie group. That group defines a one-dimensional manifold
generated by following the orbit x(t) = eGt x(0), where G is an n ? n matrix and t is a
scalar manifold coordinate. A multi-dimensional manifold can be obtained by replacing
Gt above by a linear combination of multiple generating matrices. In (Rao and Ruderman,
1999) the matrix exponential is approximated to first order by (I + Gt), and the authors
estimatePG for a simple signal undergoing translations, using as a criterion the minimiza?||2 , where x
? is a neighbor of x in the data. Note that in
tion of x,?x mint ||(I + Gt)x ? x
this model the tangent plane is a linear function of x, i.e. F1 (x) = Gx. By minimizing
the above across many pairs of examples, a good estimate of G for the artificial data was
recovered by (Rao and Ruderman, 1999). Our proposal extends this approach to multiple
dimensions and non-linear relations between x and the tangent planes. Note also the earlier
work on Tangent Distance (Simard, LeCun and Denker, 1993), in which the tangent planes
are not learned but used to build a nearest neighbor classifier that is based on the distance
between the tangent subspaces around two examples to be compared. The main advantage
of the approach proposed here over local manifold learning is that the parameters of the
tangent plane predictor can be estimated using data from very different regions of space,
thus in principle allowing to be less sensitive to all four of the problems described in 2.1,
thanks to sharing of information across these different regions.
4
Experimental Results
The objective of the experiments is to validate the proposed algorithm: does it estimate
well the true tangent planes? does it learn better than a local manifold learning algorithm?
Error Measurement In addition to visualizing the results for the low-dimensional data,
we measure performance by considering how well the algorithm learns the local tangent
distance, as measured by the normalized projection error of nearest neighbors (eq. 3). We
compare the errors of four algorithms, always on test data not used to estimate the tangent plane: (a) true analytic (using the true manifold?s tangent plane at x computed analytically), (b) tangent learning (using the neural-network tangent plane predictor F (x),
trained using the k ? d nearest neighbors in the training set of each training set example), (c) Isomap (using the tangent plane defined on Eq. 1), (d) Local PCA (using the d
principal components of the k nearest neighbors of x in the training set).
0.16
Generalization of Tangent Learning
Analytic
Local PCA
Isomap
Tangent Learning
4
0.15
0.14
3
0.13
0.12
2
0.11
1
0.1
0.09
0
0.08
0.07
?1
0.06
?2
0
1
2
3
4
5
6
7
8
9
Figure 1: Task 1 2-D data with 1-D sinusoidal manifolds: the method indeed captures the tangent planes. The small segments are the estimated tangent planes.
Red points are training examples.
1
2
3
4
5
6
7
8
9
10
10
Figure 2: Task 2 relative projection error for k-th nearest neighbor, w.r.t. k, for
compared methods (from lowest to highest at k=1: analytic, tangent learning, local PCA, Isomap). Note U-shape due to
opposing effects of curvature and noise.
Task 1 We first consider a low-dimensional but multi-manifold problem. The data {x i }
are in 2 dimensions and coming from a set of 40 1-dimensional manifolds. Each manifold is composed of 4 near points obtained from a randomly based sinus, i.e ?i ? 1..4,
xi = (a + ti , sin(a + ti ) + b, where a, b, and ti are randomly chosen. Four neighbors
were used for training both the Tangent Learning algorithm and the benchmark local nonparametric estimator (local PCA of the 4 neighbors). Figure 1 shows the training set and
the tangent planes recovered, both on the training examples and generalizing away from
the data. The neural network has 10 hidden units (chosen arbitrarily). This problem is
particularly difficult for local manifold learning, which does very poorly here: the out-ofsample relative prediction error are respectively 0.09 for the true analytic plane, 0.25 for
non-local tangent learning, and 0.81 for local PCA.
Task 2 This is a higher dimensional manifold learning problem, with 41 dimensions.
The data are generated by sampling Gaussian curves. Each curve is of the form x(i) =
2
et1 ?(?2+i/10) /t2 with i ? {0, 1, . . . , 40}. Note that the tangent vectors are not linear in x.
The manifold coordinates are t1 and t2 , sampled uniformly, respectively from (?1, 1) and
(.1, 3.1). Normal noise (standard deviation = 0.001) is added to each point. 100 example
curves were generated for training and 200 for testing. The neural network has 100 hidden
units. Figure 2 shows the relative projection error for the four methods on this task, for the
k-th nearest neighbor, for increasing values of k. First, the error decreases because of the
effect of noise (near noisy neighbors may form a high angle with the tangent plane). Then,
it increases because of the curvature of manifold (further away neighbors form a larger
angle).
Task 3 This is a high-dimensional multi-manifold task, involving digit images to which
we have applied slight rotations, in such a way as to have the knowledge of the analytic
formulation of the manifolds. There is one rotation manifold for each instance of digit from
the database, but only two examples for each manifold: one real image from the MNIST
dataset and one slightly rotated image. 1000?2 examples are used for training and 1000?2
for testing. In this context we use k = 1 nearest neighbor only and manifold dimension
is 1. The average relative projection error for the nearest neighbor is 0.27 for the analytic
tangent plane, 0.43 with tangent learning (100 hidden units), and 1.5 with Local PCA.
Here the neural network would probably overfit if trained too much (here only 100 epochs).
2
2
2
4
4
4
6
6
6
8
8
8
10
10
10
12
12
12
14
14
2
4
6
8
10
12
14
14
2
4
6
8
10
12
14
2
2
2
4
4
4
6
6
6
8
8
8
10
10
10
12
12
12
14
14
2
4
6
8
10
12
14
2
4
6
8
10
12
14
2
4
6
8
10
12
14
14
2
4
6
8
10
12
14
Figure 3: Left column: original image. Middle: applying a small amount of the predicted
rotation. Right: applying a larger amount of the predicted rotation. Top: using the estimated tangent plane predictor. Bottom: using local PCA, which is clearly much worse.
An even more interesting experiment consists in applying the above trained predictor on
novel images that come from a very different distribution but one that shares the same
manifold structure: it was applied to images of other characters that are not digits. We
have used the predicted tangent planes to follow the manifold by small steps (this is very
easy to do in the case of a one-dimensional manifold). Figure 3 shows for example on a
letter ?M? image the effect of a few such steps and a larger number of steps, both for the
neural network predictor and for the local PCA predictor.
This example illustrates the crucial point that non-local tangent plane learning is able to
generalize to truly novel cases, where local manifold learning fails.
In all the experiments we found that all the randomly initialized neural networks converged
to similarly good solutions. The number of hidden units was not optimized, although preliminary experimentation showed phenomena of over-fitting and under-fitting due to too
small or too large number hidden units was possible.
5
Conclusion
The central claim of this paper is that there are fundamental problems with non-parametric
local approaches to manifold learning, essentially due to the curse of dimensionality (at the
dimensionality of the manifold), but worsened by manifold curvature, noise, and the presence of several disjoint manifolds. To address these problems, we propose that learning
algorithms should be designed in such a way that they can share information, coming from
different regions of space, about the structure of the manifold. In this spirit we have proposed a simple learning algorithm based on predicting the tangent plane at x with a function
F (x) whose parameters are estimated based on the whole data set. Note that the same fundamental problems are present with non-parametric approaches to semi-supervised learning (e.g. as in (Szummer and Jaakkola, 2002; Chapelle, Weston and Scholkopf, 2003;
Belkin and Niyogi, 2003; Zhu, Ghahramani and Lafferty, 2003)), which rely on proper
estimation of the manifold in order to propagate label information.
Future work should investigate how to better handle the curvature problem, e.g. by follow-
ing the manifold (using the local tangent estimates), to estimate a manifold-following path
between pairs of neighboring examples. The algorithm can also be extended in a straightforward way to obtain a Gaussian mixture or a mixture of factor analyzers (with the factors
or the principal eigenvectors of the Gaussian centered at x obtained from F (x)). This view
can also provide an alternative criterion to optimize F (x) (the local log-likelihood of such
a Gaussian). This criterion also tells us how to estimate the missing information (the variances along the eigenvector directions). Since we can estimate F (x) everywhere, a more
ambitious view would consider the density as a ?continuous? mixture of Gaussians (with
an infinitesimal component located everywhere in space).
Acknowledgments
The authors would like to thank the following funding organizations for support: NSERC,
MITACS, IRIS, and the Canada Research Chairs.
References
Belkin, M. and Niyogi, P. (2003). Using manifold structure for partially labeled classification. In
Becker, S., Thrun, S., and Obermayer, K., editors, Advances in Neural Information Processing
Systems 15, Cambridge, MA. MIT Press.
Bengio, Y., Delalleau, O., Le Roux, N., Paiement, J.-F., Vincent, P., and Ouimet, M. (2004). Learning
eigenfunctions links spectral embedding and kernel PCA. Neural Computation, 16(10):2197?
2219.
Brand, M. (2003). Charting a manifold. In Becker, S., Thrun, S., and Obermayer, K., editors,
Advances in Neural Information Processing Systems 15. MIT Press.
Chapelle, O., Weston, J., and Scholkopf, B. (2003). Cluster kernels for semi-supervised learning. In
Becker, S., Thrun, S., and Obermayer, K., editors, Advances in Neural Information Processing
Systems 15, Cambridge, MA. MIT Press.
Ghahramani, Z. and Hinton, G. (1996). The EM algorithm for mixtures of factor analyzers. Technical
Report CRG-TR-96-1, Dpt. of Comp. Sci., Univ. of Toronto.
Rao, R. and Ruderman, D. (1999). Learning lie groups for invariant visual perception. In Kearns,
M., Solla, S., and Cohn, D., editors, Advances in Neural Information Processing Systems 11,
pages 810?816. MIT Press, Cambridge, MA.
Roweis, S. and Saul, L. (2000). Nonlinear dimensionality reduction by locally linear embedding.
Science, 290(5500):2323?2326.
Sch?
olkopf, B., Smola, A., and M?
uller, K.-R. (1998). Nonlinear component analysis as a kernel
eigenvalue problem. Neural Computation, 10:1299?1319.
Simard, P., LeCun, Y., and Denker, J. (1993). Efficient pattern recognition using a new transformation
distance. In Giles, C., Hanson, S., and Cowan, J., editors, Advances in Neural Information
Processing Systems 5, pages 50?58, Denver, CO. Morgan Kaufmann, San Mateo.
Szummer, M. and Jaakkola, T. (2002). Partially labeled classification with markov random walks.
In Dietterich, T., Becker, S., and Ghahramani, Z., editors, Advances in Neural Information
Processing Systems 14, Cambridge, MA. MIT Press.
Teh, Y. W. and Roweis, S. (2003). Automatic alignment of local representations. In Becker, S.,
Thrun, S., and Obermayer, K., editors, Advances in Neural Information Processing Systems 15.
MIT Press.
Tenenbaum, J., de Silva, V., and Langford, J. (2000). A global geometric framework for nonlinear
dimensionality reduction. Science, 290(5500):2319?2323.
Tipping, M. and Bishop, C. (1999). Mixtures of probabilistic principal component analysers. Neural
Computation, 11(2):443?482.
Vincent, P. and Bengio, Y. (2003). Manifold parzen windows. In Becker, S., Thrun, S., and Obermayer, K., editors, Advances in Neural Information Processing Systems 15, Cambridge, MA.
MIT Press.
Zhu, X., Ghahramani, Z., and Lafferty, J. (2003). Semi-supervised learning using gaussian fields and
harmonic functions. In ICML?2003.
| 2647 |@word illustrating:1 middle:1 propagate:1 decomposition:1 covariance:6 tr:1 reduction:3 egt:1 existing:1 recovered:2 yet:1 must:1 written:1 shape:2 analytic:7 designed:1 selected:1 plane:30 xk:4 characterization:1 provides:1 toronto:1 successive:1 gx:1 along:1 constructed:2 scholkopf:2 consists:1 combine:1 fitting:2 indeed:2 nor:1 multi:3 globally:1 little:1 curse:3 window:2 considering:1 increasing:2 discover:1 underlying:2 estimating:2 lowest:1 what:1 kind:2 minimizes:2 eigenvector:1 transformation:3 guarantee:1 ti:3 exactly:2 universit:1 scaled:1 classifier:1 unit:5 t1:1 understood:1 local:56 consequence:1 matrixvalued:1 path:1 interpolation:1 might:1 mateo:1 suggests:1 co:1 acknowledgment:1 responsible:1 lecun:2 testing:2 union:1 practice:1 nystro:1 digit:4 significantly:1 projection:8 get:1 close:1 selection:1 context:3 impossible:1 applying:3 optimize:1 equivalent:1 demonstrated:1 center:2 missing:1 go:1 straightforward:1 resolution:2 qc:1 roux:1 estimator:1 spanned:2 embedding:12 handle:1 notion:2 coordinate:11 imagine:2 canpbe:1 target:1 associate:1 element:2 approximated:3 particularly:1 located:1 recognition:1 predicts:1 database:1 labeled:2 observed:2 role:1 bottom:1 capture:2 region:5 solla:1 decrease:1 highest:1 pd:1 geodesic:1 trained:5 depend:2 solving:1 segment:1 minimiza:1 wtj:12 basis:3 k0:1 represented:1 worsened:1 alphabet:1 train:1 univ:1 artificial:1 tell:1 analyser:1 neighborhood:1 exhaustive:1 whose:1 larger:3 say:1 delalleau:1 niyogi:4 h3c:1 noisy:2 advantage:3 eigenvalue:1 propose:2 interaction:1 coming:2 neighboring:1 combining:1 poorly:1 roweis:7 moved:1 validate:1 olkopf:1 convergence:1 double:1 cluster:1 generating:1 rotated:1 object:4 depending:1 montreal:1 measured:1 nearest:10 eq:3 strong:2 predicted:4 involves:1 come:1 direction:8 closely:2 downtown:1 stochastic:2 centered:1 f1:1 generalization:2 preliminary:1 enumerated:1 crg:1 lying:1 around:12 normal:1 claim:3 vki:2 estimation:1 label:1 coordination:1 sensitive:1 et1:1 create:1 uller:3 minimization:2 mit:7 clearly:3 j7:1 gaussian:11 always:1 pn:1 avoid:1 jaakkola:2 vk:3 properly:2 likelihood:1 mainly:1 contrast:2 dependent:5 hidden:6 relation:3 going:2 pixel:2 issue:1 overall:1 classification:2 field:1 once:1 sampling:1 unsupervised:1 icml:1 future:1 yoshua:1 t2:2 fundamentally:1 report:1 belkin:4 few:2 randomly:3 composed:1 n1:1 opposing:1 attempt:1 montr:1 organization:1 investigate:1 alignment:1 weakness:2 mixture:13 truly:1 edge:3 closer:1 necessary:2 pancake:3 initialized:1 orbit:1 walk:1 increased:1 eal:1 modeling:1 instance:3 rao:4 earlier:1 ar:1 column:1 giles:1 ordinary:1 deviation:1 vertex:1 entry:1 expects:1 predictor:6 eigenmaps:3 too:3 characterize:3 sv:1 thanks:1 density:6 fundamental:3 probabilistic:2 parzen:2 quickly:1 central:2 choose:1 possibly:3 worse:1 ek:3 derivative:1 simard:2 sinusoidal:1 de:4 depends:1 tion:1 view:3 lot:1 red:1 complicated:1 minimize:2 square:1 variance:1 characteristic:1 kaufmann:1 correspond:1 yield:1 generalize:5 lkopf:2 handwritten:1 vincent:4 basically:2 lighting:1 comp:1 classified:1 converged:1 sharing:1 infinitesimal:1 failure:1 obvious:1 associated:1 sampled:1 dataset:2 knowledge:1 dimensionality:11 higher:1 tipping:3 supervised:4 follow:2 formulation:1 done:3 box:1 mitacs:1 just:1 smola:3 langford:3 overfit:1 ruderman:4 replacing:1 monperrus:1 cohn:1 nonlinear:3 defines:1 effect:4 dietterich:1 normalized:1 true:5 isomap:8 hence:1 analytically:3 i2:1 illustrated:1 visualizing:1 sin:1 iris:1 criterion:10 silva:3 image:16 harmonic:1 novel:2 funding:1 umontreal:1 common:1 rotation:6 physical:1 denver:1 exponentially:2 slight:1 measurement:1 cambridge:5 framed:1 rd:3 automatic:1 similarly:1 analyzer:3 dj:1 had:1 chapelle:2 gt:3 something:2 curvature:10 recent:2 showed:1 mint:1 compound:1 store:1 arbitrarily:1 seen:2 minimum:1 greater:1 morgan:1 determine:1 signal:1 semi:3 branch:1 multiple:2 ing:1 technical:1 match:2 calculation:1 devised:1 laplacian:3 prediction:3 involving:1 essentially:3 vision:1 sinus:2 kernel:19 represent:1 normalization:1 proposal:1 addition:1 separately:1 grow:2 singular:1 crucial:1 sch:3 extra:1 probably:1 eigenfunctions:1 cowan:1 lafferty:2 spirit:1 call:1 near:12 presence:3 bengio:9 easy:2 enough:1 variety:2 xj:22 vik:1 pca:17 allocate:1 becker:6 suffer:1 cause:1 generally:1 eigenvectors:4 amount:6 nonparametric:1 locally:6 tenenbaum:3 simplest:1 specifies:1 estimated:7 disjoint:1 per:3 paiement:1 group:4 four:5 graph:3 year:1 orient:1 angle:3 parameterized:1 letter:1 everywhere:2 striking:1 extends:1 saying:1 almost:1 patch:7 scaling:1 layer:1 constraint:1 argument:1 span:2 min:1 chair:1 martin:1 according:1 alternate:1 combination:5 kd:6 smaller:1 slightly:2 across:2 character:3 em:1 wi:1 invariant:1 equation:2 precomputed:1 fail:1 needed:5 ouimet:1 gaussians:5 experimentation:1 apply:2 denker:2 away:3 spectral:3 robustly:1 alternative:1 original:2 dpt:1 denotes:1 clustering:2 include:1 top:1 giving:1 ghahramani:6 build:2 move:1 objective:1 already:1 added:1 parametric:5 obermayer:5 gradient:4 subspace:6 distance:6 thank:1 link:1 thrun:5 sci:1 manifold:103 collected:1 iro:2 charting:2 index:1 ratio:1 minimizing:2 optionally:1 difficult:3 mostly:4 potentially:1 holding:1 implementation:1 proper:1 ambitious:1 perform:1 teh:4 allowing:1 observation:1 hurtful:1 markov:1 arc:1 benchmark:1 descent:2 optional:1 hinton:3 extended:1 rn:1 canada:2 pair:4 required:2 specified:1 optimized:2 hanson:1 coherent:1 learned:2 address:2 able:3 below:2 usually:1 ofsample:1 perception:1 pattern:1 overlap:1 rely:1 predicting:1 zhu:2 prior:1 epoch:1 geometric:1 tangent:49 relative:5 sk2:1 interesting:1 generator:1 principle:1 editor:8 share:4 translation:7 row:1 free:1 lle:3 saul:3 neighbor:32 characterizing:1 curve:4 dimension:7 xn:5 gram:1 world:1 author:2 adaptive:1 san:1 far:5 approximate:1 implicitly:1 global:3 xi:17 continuous:1 sk:1 learn:3 ca:1 interpolating:1 diag:1 did:1 main:1 whole:3 noise:10 x1:1 fails:2 position:2 exponential:1 bengioy:1 lie:4 learns:1 formula:1 down:1 bishop:3 xt:22 showing:1 undergoing:1 list:1 derives:1 intrinsic:1 mnist:1 flattened:1 illustrates:1 nk:2 locality:1 generalizing:1 simply:4 explore:2 visual:1 expressed:1 nserc:1 partially:2 scalar:1 corresponds:1 ma:5 weston:2 shared:1 bors:1 change:3 included:1 determined:1 uniformly:1 principal:9 kearns:1 total:1 experimental:1 brand:5 meaningful:1 support:1 szummer:2 dept:1 phenomenon:3 |
1,814 | 2,648 | Semi-Markov Conditional Random Fields for
Information Extraction
Sunita Sarawagi
Indian Institute of Technology
Bombay, India
[email protected]
William W. Cohen
Center for Automated Learning & Discovery
Carnegie Mellon University
[email protected]
Abstract
We describe semi-Markov conditional random fields (semi-CRFs), a conditionally trained version of semi-Markov chains. Intuitively, a semiCRF on an input sequence x outputs a ?segmentation? of x, in which
labels are assigned to segments (i.e., subsequences) of x rather than to
individual elements xi of x. Importantly, features for semi-CRFs can
measure properties of segments, and transitions within a segment can
be non-Markovian. In spite of this additional power, exact learning and
inference algorithms for semi-CRFs are polynomial-time?often only a
small constant factor slower than conventional CRFs. In experiments
on five named entity recognition problems, semi-CRFs generally outperform conventional CRFs.
1
Introduction
Conditional random fields (CRFs) are a recently-introduced formalism [12] for representing a conditional model Pr(y|x), where both x and y have non-trivial structure (often
sequential). Here we introduce a generalization of sequential CRFs called semi-Markov
conditional random fields (or semi-CRFs). Recall that semi-Markov chain models extend
hidden Markov models (HMMs) by allowing each state si to persist for a non-unit length of
time di . After this time has elapsed, the system will transition to a new state s0 , which depends only on si ; however, during the ?segment? of time between i and i + di , the behavior
of the system may be non-Markovian. Generative semi-Markov models are fairly common
in certain applications of statistics [8, 9], and are also used in reinforcement learning to
model hierarchical Markov decision processes [19].
Semi-CRFs are a conditionally trained version of semi-Markov chains. In this paper, we
present inference and learning methods for semi-CRFs. We also argue that segments often
have a clear intuitive meaning, and hence semi-CRFs are more natural than conventional
CRFs. We focus here on named entity recognition (NER), in which a segment corresponds
to an extracted entity; however, similar arguments might be made for several other tasks,
such as gene-finding [11] or NP-chunking [16].
In NER, a semi-Markov formulation allows one to easily construct entity-level features
(such as ?entity length? and ?similarity to other known entities?) which cannot be easily
encoded in CRFs. Experiments on five different NER problems show that semi-CRFs often
outperform conventional CRFs.
2
2.1
CRFs and Semi-CRFs
Definitions
A CRF models Pr(y|x) using a Markov random field, with nodes corresponding to elements of the structured object y, and potential functions that are conditional on (features
of) x. Learning is performed by setting parameters to maximize the likelihood of a set of
(x, y) pairs given as training data. One common use of CRFs is for sequential learning
problems like NP chunking [16], POS tagging [12], and NER [15]. For these problems
the Markov field is a chain, and y is a linear sequence of labels from a fixed set Y. For instance, in the NER application, x might be a sequence of words, and y might be a sequence
in {I, O}|x| , where yi = I indicates ?word xi is inside a name? and yi = O indicates the
opposite.
Assume a vector f of local feature functions f = hf 1 , . . . , f K i, each of which maps a pair
(x, y) and an index i to a measurement f k (i, x, y) ? R. Let f (i, x, y) be the vector of these
P|x|
measurements, and let F(x, y) = i f (i, x, y). For example, in NER, the components
of f might include the measurement f 13 (i, x, y) = [[xi is capitalized]] ? [[yi = I]], where
the indicator function [[c]] = 1 if c if true and zero otherwise; this implies that F 13 (x, y)
would be the number of capitalized words xi paired with the label I. Following previous
work [12, 16] we will define a conditional random field (CRF) to be an estimator of the
form
1 W?F(x,y)
e
(1)
Pr(y|x, W) =
Z(x)
P W?F(x,y0 )
where W is a weight vector over the components of F, and Z(x) = y0 e
.
To extend this to the semi-Markov case, let s = hs1 , . . . , sp i denote a segmentation of x,
where segment sj = htj , uj , yj i consists of a start position tj , an end position uj , and a
label yj ? Y . Conceptually, a segment means that the tag yj is given to all xi ?s between
i = tj and i = uj , inclusive. We assume segments have positive length, and completely
cover the sequence 1 . . . |x| without overlapping: that is, that tj and uj always satisfy
t1 = 1, up = |x|, 1 ? tj ? uj ? |x|, and tj+1 = uj + 1. For NER, a valid segmentation
of the sentence ?I went skiing with Fernando Pereira in British Columbia? might be s =
h(1, 1, O), (2, 2, O), (3, 3, O), (4, 4, O), (5, 6, I), (7, 7, O), (8, 9, I)i, corresponding to the
label sequence y = hO, O, O, O, I, I, O, I, Ii.
We now assume a vector g of segment feature functions g = hg 1 , . . . , g K i, each of
which maps a triple (j, x, s) to a measurement g k (j, x, s) ? R, and define G(x, s) =
P|s|
j g(j, x, s). We also make a restriction on the features, analogous to the usual Markovian assumption made in CRFs, and assume that every component g k of g is a function
only of x, sj , and the label yj?1 associated with the preceding segment sj?1 . In other
words, we assume that every g k (j, x, s) can be rewritten as
g k (j, x, s) = g 0k (yj , yj?1 , x, tj , uj )
(2)
for an appropriately defined g 0k . In the rest of the paper, we will drop the g 0 notation and
use g for both versions of the segment-level feature functions. A semi-CRF is then an
estimator of the form
1 W?G(x,s)
Pr(s|x, W) =
e
(3)
Z(x)
P W?G(x,s0 )
where again W is a weight vector for G and Z(x) = s0 e
.
2.2
An efficient inference algorithm
The inference problem for a semi-CRF is defined as follows: given W and x, find the
best segmentation, argmax s Pr(s|x, W), where Pr(s|x, W) is defined by Equation 3. An
efficient inference algorithm is suggested by Equation 2, which implies that
X
argmax s Pr(s|x, W) = argmax s W ? G(x, s) = argmax s W ?
g(yj , yj?1 , x, tj , uj )
j
Let L be an upper bound on segment length. Let si:y denote the set of all partial segmentations starting from 1 (the first index of the sequence) to i, such that the last segment has
the label y and ending position i. Let Vx,g,W (i, y) denote the largest value of W ? G(x, s0 )
for any s0 ? si:y . Omitting the subscripts, the following recursive calculation implements
a semi-Markov analog of the usual Viterbi algorithm:
(
maxy0 ,d=1...L V (i ? d, y 0 ) + W ? g(y, y 0 , x, i ? d + 1, i) if i > 0
0
if i = 0
V (i, y) =
(4)
??
if i < 0
The best segmentation then corresponds to the path traced by maxy V (|x|, y).
2.3
Semi-Markov CRFs vs order-L CRFs
Since conventional CRFs need not maximize over possible segment lengths d, inference
for semi-CRFs is more expensive. However, Equation 4 shows that the additional cost is
only linear in L. For NER, a reasonable value of L might be four or five.1 Since in the
worst case L ? |x|, the semi-Markov Viterbi algorithm is always polynomial, even when
L is unbounded.
For fixed L, it can be shown that semi-CRFs are no more expressive than order-L CRFs.
For order-L CRFs, however the additional computational cost is exponential in L. The
difference is that semi-CRFs only consider sequences in which the same label is assigned
to all L positions, rather than all |Y|L length-L sequences. This is a useful restriction, as it
leads to faster inference.
Semi-CRFs are also a natural restriction, as it is often convenient to express features in
terms of segments. As an example, let dj denote the length of a segment, and let ?
be the average length of all segments with label I. Now consider the segment feature
g k1 (j, x, s) = (dj ? ?)2 ? [[yj = I]]. After training, the contribution of this feature toward
2
Pr(s|x) associated with a length-d entity will be proportional to ewk ?(d??) ?i.e., it allows
the learner to model a Gaussian distribution of entity lengths.
An exponential model for lengths could be implemented with the feature g k2 (j, x, y) =
dj ? [[yj = I]]. In contrast to the Gaussian-length feature above, g k2 is ?equivalent to? a
local feature function f (i,P
x, y) = [[yi = I]], in
Pthe following sense: for every triple x, y, s,
where y is the tags for s, j g k2 (j, x, s) = i f (i, s, y). Thus a semi-CRF model based
on the single feature g k2 could also be represented by a conventional CRF.
In general, a semi-CRF model can be factorized in terms of an equivalent order-1 CRF
model if and only if the sum of the segment features can be rewritten as a sum of local
features. Thus the degree to which semi-CRFs are non-Markovian depends on the feature
set.
2.4
Learning algorithm
During training the goal is to maximize log-likelihood over a given training set T =
{(x` , s` )}N
`=1 . Following the notation of Sha and Pereira [16], we express the loglikelihood over the training sequences as
X
X
L(W) =
log Pr(s` |x` , W) =
(W ? G(x` , s` ) ? log ZW (x` ))
(5)
`
1
`
Assuming that non-entity words are placed in unit-length segments, as we do below.
We wish to find a W that maximizes L(W). Equation 5 is convex, and can thus be maximized by gradient ascent, or one of many related methods. (In our implementation we use
a limited-memory quasi-Newton method [13, 14].) The gradient of L(W) is the following:
P
0
W?G(x` ,s0 )
X
0 G(s , x` )e
?L(W) =
G(x` , s` ) ? s
(6)
ZW (x` )
`
X
=
G(x` , s` ) ? EPr(s0 |W) G(x` , s0 )
(7)
`
The first set of terms are easy to compute. However, to compute the the normalizer, ZW (x` ), and the expected value of the features under the current weight vector,
EPr(s0 |W) G(x` , s0 ), we must use the Markov property of G to construct a dynamic programming algorithm, similar for forward-backward. We thus define ?(i, y) as the value of
P
W?G(s0 ,x)
where again si:y denotes all segmentations from 1 to i ending at i and
s0 ?si:y e
labeled y. For i > 0, this can be expressed recursively as
?(i, y) =
L X
X
0
?(i ? d, y 0 )eW?g(y,y ,x,i?d+1,i)
d=1 y 0 ?Y
with the base cases defined as ?(0,
Py) = 1 and ?(i, y) = 0 for i < 0. The value of ZW (x)
can then be written as ZW (x) = y ?(|x|, y).
P
0
A similar approach can be used to compute the expectation s0 G(x` , s0 )eW?G(x` ,s ) .
For the k-th component of G, let ? k (i, y) be the value of the sum
P
k 0
W?G(x` ,s0 )
, restricted to the part of the segmentation ending at
s0 ?si:y G (s , x` )e
position i. The following recursion2 can then be used to compute ? k (i, y):
? k (i, y) =
PL P
0
(i?d, y 0 ) + ?(i?d, y 0 )g k (y, y 0 , x, i?d + 1, i))eW?g(y,y ,x,i?d+1,i)
P
Finally we let EPr(s0 |W) Gk (s0 , x) = ZW1(x) y ? k (|x|, y).
d=1
3
3.1
y 0 ?Y (?
k
Experiments with NER data
Baseline algorithms and datasets
In our experiments, we trained semi-CRFs to mark entity segments with the label I, and
put non-entity words into unit-length segments with label O. We compared this with two
versions of CRFs. The first version, which we call CRF/1, labels words inside and outside
entities with I and O, respectively. The second version, called CRF/4, replaces the I tag
with four tags B, E, C, and U , which depend on where the word appears in an entity [2].
We compared the algorithms on five NER problems, associated with three different corpora.
The Address corpus contains 4,226 words, and consists of 395 home addresses of students
in a major university in India [1]. We considered extraction of city names and state names
from this corpus. The Jobs corpus contains 73,330 words, and consists of 300 computerrelated job postings [4]. We considered extraction of company names and job titles. The
18,121-word Email corpus contains 216 email messages taken from the CSPACE email
corpus [10], which is mail associated with a 14-week, 277-person management game. Here
we considered extraction of person names.
2
As in the forward-backward algorithm for chain CRFs [16], space requirements here can be
reduced from M L|Y| to M |Y|, where M is the length of the sequence, by pre-computing an appropriate set of ? values.
3.2
Features
As features for CRF, we used indicators for specific words at location i, or locations within
three words of i. Following previous NER work [7]), we also used indicators for capitalization/letter patterns (such as ?Aa+? for a capitalized word, or ?D? for a single-digit
number).
As features for semi-CRFs, we used the same set of word-level features, as well their
logical extensions to segments. Specifically, we used indicators for the phrase inside a
segment and the capitalization pattern inside a segment, as well as indicators for words
and capitalization patterns in 3-word windows before and after the segment. We also used
indicators for each segment length (d = 1, . . . , L), and combined all word-level features
with indicators for the beginning and end of a segment.
To exploit more of the power of semi-CRFs, we also implemented a number of dictionaryderived features, each of which was based on different dictionary D and similarity function
sim. Letting xsj denote the subsequence hxtj . . . xuj i, a dictionary feature is defined as
g D,sim (j, x, s) = argmax u?D sim(xsj , u)?i.e., the distance from the word sequence xsj
to the closest element in D.
For each of the extraction problems, we assembled one external dictionary of strings, which
were similar (but not identical) to the entity names in the documents. For instance, for
city names in the Address data, we used a web page listing cities in India. Due to variations in the way entity names are written, rote matching these dictionaries to the data
gives relatively low F1 values, ranging from 22% (for the job-title extraction task) to 57%
(for the person-name task). We used three different similarity metrics (Jaccard, TFIDF,
and JaroWinkler) which are known to work well for name-matching in data integration
tasks [5]. All of the distance metrics are non-Markovian?i.e., the distance-based segment
features cannot be decomposed into sums of local features. More detail on the distance
metrics, feature sets, and datasets above can be found elsewhere [6].
We also extended the semi-CRF algorithm to construct, on the fly, an internal segment
dictionary of segments labeled as entities in the training data. To make measurements on
training data similar to those on test data, when finding the closest neighbor of xsj in the
internal dictionary, we excluded all strings formed from x, thus excluding matches of xsj to
itself (or subsequences of itself). This feature could be viewed as a sort of nearest-neighbor
classifier; in this interpretation the semi-CRF is performing a sort of bi-level stacking [21].
For completeness in the experiments, we also evaluated local versions of the dictionary
features. Specifically, we constructed dictionary features of the form f D,sim (i, x, y) =
argmax u?D sim(xi , u), where D is either the external dictionary used above, or an internal
word dictionary formed from all words contained in entities. As before, words in x were
excluded in finding near neighbors to xi .
3.3
Results and Discussion
We evaluated F1-measure performance3 of CRF/1, CRF/4, and semi-CRFs, with and without internal and external dictionaries. A detailed tabulation of the results are shown in Table 1, and Figure 1 shows F1 values plotted against training set size for a subset of three of
the tasks, and four of the learning methods. In each experiment performance was averaged
over seven runs, and evaluation was performed on a hold-out set of 30% of the documents.
In the table the learners are trained with 10% of the available data?as the curves show,
performance differences are often smaller with more training data. Gaussian priors were
used for all algorithms, and for semi-CRFs, a fixed value of L was chosen for each dataset
based on observed entity lengths. This ranged between 4 and 6 for the different datasets.
In the baseline configuration in which no dictionary features are used, semi-CRFs perform
3
F1 is defined as 2*precision*recall/(precision+recall.)
Address_City
Email_Person
90
90
85
85
F1 span accuracy
F1 span accuracy
F1 span accuracy
Address_State
100
90
80
70
60
50
40
CRF/4
30
SemiCRF+int
20
CRF/4+dict
10
SemiCRF+int+dict
0
0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
Fraction of available training data
80
75
CRF/4
SemiCRF+int
CRF/4+dict
SemiCRF+int+dict
70
65
0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
Fraction of available training data
80
75
70
CRF/4
SemiCRF+int
CRF/4+dict
SemiCRF+int+dict
65
0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
Fraction of available training data
Figure 1: F1 as a function of training set size. Algorithms marked with ?+dict? include external
dictionary features, and algorithms marked with ?+int? include internal dictionary features. We do
not use internal dictionary features for CRF/4 since they lead to reduced accuracy.
baseline
F1
CRF/1
state
title
person
city
company
CRF/4
state
title
person
city
company
semi-CRF
state
title
person
city
company
+internal dict
F1 ?base
+external dict
F1 ?base
+both dictionaries
F1 ?base ?extern
20.8
28.5
67.6
70.3
51.4
44.5
3.8
48.0
60.0
16.5
113.9
-86.7
-29.0
-14.7
-67.9
69.2
38.6
81.4
80.4
55.3
232.7
35.4
20.4
14.4
7.6
55.2
19.9
64.7
69.8
15.6
165.4
-30.2
-4.3
-0.7
-69.6
-67.3
-65.6
-24.7
-15.1
-77.2
15.0
23.7
70.9
73.2
54.8
25.4
7.9
64.5
70.6
20.6
69.3
-66.7
-9.0
-3.6
-62.4
46.8
36.4
82.5
80.8
61.2
212.0
53.6
16.4
10.4
11.7
43.1
14.6
74.8
76.3
25.1
187.3
-38.4
5.5
4.2
-54.2
-24.7
-92.0
-10.9
-6.1
-65.9
25.6
33.8
72.2
75.9
60.2
35.5
37.5
74.8
75.3
59.7
38.7
10.9
3.6
-0.8
-0.8
62.7
41.1
82.8
84.0
60.9
144.9
21.5
14.7
10.7
1.2
65.2
40.2
83.7
83.6
60.9
154.7
18.9
15.9
10.1
1.2
9.8
-2.5
1.2
-0.5
0.0
Table 1: Comparing various methods on five IE tasks, with and without dictionary features. The
column ?base is percentage change in F1 values relative to the baseline. The column ?extern is is
change relative to using only external-dictionary features.
best on all five of the tasks. When internal dictionary features are used, the performance
of semi-CRFs is often improved, and never degraded by more than 2.5%. However, the
less-natural local version of these features often leads to substantial performance losses for
CRF/1 and CRF/4. Semi-CRFs perform best on nine of the ten task variants for which
internal dictionaries were used. The external-dictionary features are helpful to all the algorithms. Semi-CRFs performs best on three of five tasks in which only external dictionaries
were used.
Overall, semi-CRF performs quite well. If we consider the tasks with and without external
dictionary features as separate ?conditions?, then semi-CRFs using all available information4 outperform both CRF variants on eight of ten ?conditions?.
We also compared semi-CRF to order-L CRFs, with various values of L.5 In Table 2 we
show the result for L = 1, L = 2, and L = 3, compared to semi-CRF. For these tasks, the
performance of CRF/4 and CRF/1 does not seem to improve much by simply increasing
4
I.e., the both-dictionary version when external dictionaries are available, and the internaldictionary only version otherwise.
5
Order-L CRFs were implemented by replacing the label set Y with Y L . We limited experiments
to L ? 3 for computational reasons.
Address State
Address City
Email persons
L=1
20.8
70.3
67.6
CRF/1
L=2
20.1
71.0
63.7
L=3
19.2
71.2
66.7
L=1
15.0
73.2
70.9
CRF/4
L=2
16.4
73.9
70.7
semi-CRF
L=3
16.4
73.7
70.4
25.6
75.9
72.2
Table 2: F1 values for different order CRFs
order.
4
Related work
Semi-CRFs are similar to nested HMMs [1], which can also be trained discriminitively [17]. The primary difference is that the ?inner model? for semi-CRFs is of short,
uniformly-labeled segments with non-Markovian properties, while nested HMMs allow
longer, diversely-labeled, Markovian ?segments?.
Discriminative learning methods can be used for conditional random fields with architectures more complex than chains (e.g., [20, 18]), and one of these methods has also been
applied to NER [3]. Further, by creating a random variable for each possible segment, one
can learn models strictly more expressive than the semi-Markov models described here.
However, for these methods, inference is not tractable, and hence approximations must be
made in training and classification. An interesting question for future research is whether
the tractible extension to CRF inference considered here can can be used to improve inference methods for more expressive models.
In recent prior work [6], we investigated semi-Markov learning methods for NER based
on a voted perceptron training algorithm [7]. The voted perceptron has some advantages
in ease of implementation, and efficiency. (In particular, the voted perceptron algorithm
is more stable numerically, as it does not need to compute a partition function. ) However, semi-CRFs perform somewhat better, on average, than our perceptron-based learning
algorithm. Probabilistically-grounded approaches like CRFs also are preferable to marginbased approaches like the voted perceptron in certain settings, e.g., when it is necessary to
estimate confidences in a classification.
5
Concluding Remarks
Semi-CRFs are a tractible extension of CRFs that offer much of the power of higher-order
models without the associated computational cost. A major advantage of semi-CRFs is that
they allow features which measure properties of segments, rather than individual elements.
For applications like NER and gene-finding [11], these features can be quite natural.
Appendix
An implementation of semi-CRFs is available at http://crf.sourceforge.net, and a NER
package using this package is available on http://minorthird.sourceforge.net.
References
[1] V. R. Borkar, K. Deshmukh, and S. Sarawagi. Automatic text segmentation for extracting
structured records. In Proc. ACM SIGMOD International Conf. on Management of Data, Santa
Barabara,USA, 2001.
[2] A. Borthwick, J. Sterling, E. Agichtein, and R. Grishman. Exploiting diverse knowledge sources
via maximum entropy in named entity recognition. In Sixth Workshop on Very Large Corpora
New Brunswick, New Jersey. Association for Computational Linguistics., 1998.
[3] R. Bunescu and R. J. Mooney. Relational markov networks for collective information extraction. In Proceedings of the ICML-2004 Workshop on Statistical Relational Learning (SRL2004), Banff, Canada, July 2004.
[4] M. E. Califf and R. J. Mooney. Bottom-up relational learning of pattern matching rules for
information extraction. Journal of Machine Learning Research, 4:177?210, 2003.
[5] W. W. Cohen, P. Ravikumar, and S. E. Fienberg. A comparison of string distance metrics for
name-matching tasks. In Proceedings of the IJCAI-2003 Workshop on Information Integration
on the Web (IIWeb-03), 2003.
[6] W. W. Cohen and S. Sarawagi. Exploiting dictionaries in named entity extraction: Combining
semi-markov extraction processes and data integration methods. In Proceedings of the Tenth
ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2004.
[7] M. Collins. Discriminative training methods for hidden Markov models: Theory and experiments with perceptron algorithms. In Empirical Methods in Natural Language Processing
(EMNLP), 2002.
[8] X. Ge. Segmental Semi-Markov Models and Applications to Sequence Analysis. PhD thesis,
University of California, Irvine, December 2002.
[9] J. Janssen and N. Limnios. Semi-Markov Models and Applications. Kluwer Academic, 1999.
[10] R. E. Kraut, S. R. Fussell, F. J. Lerch, and J. A. Espinosa. Coordination in teams: evidence from
a simulated management game. To appear in the Journal of Organizational Behavior, 2005.
[11] A. Krogh. Gene finding: putting the parts together. In M. J. Bishop, editor, Guide to Human
Genome Computing, pages 261?274. Academic Press, 2nd edition, 1998.
[12] J. Lafferty, A. McCallum, and F. Pereira. Conditional random fields: Probabilistic models for
segmenting and labeling sequence data. In Proceedings of the International Conference on
Machine Learning (ICML-2001), Williams, MA, 2001.
[13] D. C. Liu and J. Nocedal. On the limited memory BFGS method for large-scale optimization.
Mathematic Programming, 45:503?528, 1989.
[14] R. Malouf. A comparison of algorithms for maximum entropy parameter estimation. In Proceedings of The Sixth Conference on Natural Language Learning (CoNLL-2002), pages 49?55,
2002.
[15] A. McCallum and W. Li. Early results for named entity recognition with conditional random
fields, feature induction and web-enhanced lexicons. In Proceedings of The Seventh Conference
on Natural Language Learning (CoNLL-2003), Edmonton, Canada, 2003.
[16] F. Sha and F. Pereira. Shallow parsing with conditional random fields. In Proceedings of HLTNAACL, 2003.
[17] M. Skounakis, M. Craven, and S. Ray. Hierarchical hidden Markov models for information
extraction. In Proceedings of the 18th International Joint Conference on Artificial Intelligence,
Acapulco, Mexico. Morgan Kaufmann., 2003.
[18] C. Sutton, K. Rohanimanesh, and A. McCallum. Dynamic conditional random fields: Factorized probabilistic models for labeling and segmenting sequence data. In ICML, 2004.
[19] R. Sutton, D. Precup, and S. Singh. Between MDPs and semi-MDPs: A framework for temporal
abstraction in reinforcement learning. Artificial Intelligence, 112:181?211, 1999.
[20] B. Taskar, P. Abbeel, and D. Koller. Discriminative probabilistic models for relational data. In
Proceedings of Eighteenth Conference on Uncertainty in Artificial Intelligence (UAI02), Edmonton, Canada, 2002.
[21] D. H. Wolpert. Stacked generalization. Neural Networks, 5:241?259, 1992.
| 2648 |@word version:10 polynomial:2 nd:1 recursively:1 configuration:1 contains:3 liu:1 document:2 current:1 comparing:1 si:7 must:2 parsing:1 written:2 partition:1 rote:1 drop:1 v:1 generative:1 intelligence:3 mccallum:3 beginning:1 short:1 record:1 completeness:1 node:1 location:2 banff:1 lexicon:1 five:7 unbounded:1 constructed:1 consists:3 ray:1 inside:4 introduce:1 tagging:1 expected:1 behavior:2 decomposed:1 company:4 window:1 increasing:1 notation:2 maximizes:1 factorized:2 string:3 htj:1 finding:5 tractible:2 temporal:1 every:3 preferable:1 k2:4 classifier:1 unit:3 appear:1 segmenting:2 positive:1 ner:15 t1:1 local:6 before:2 sutton:2 subscript:1 path:1 might:6 hmms:3 limited:3 ease:1 bi:1 averaged:1 yj:10 recursive:1 epr:3 implement:1 sarawagi:3 digit:1 empirical:1 convenient:1 matching:4 word:22 pre:1 confidence:1 spite:1 cannot:2 put:1 py:1 restriction:3 conventional:6 map:2 equivalent:2 center:1 crfs:53 eighteenth:1 williams:1 starting:1 convex:1 estimator:2 rule:1 importantly:1 variation:1 analogous:1 enhanced:1 exact:1 programming:2 element:4 recognition:4 expensive:1 persist:1 labeled:4 observed:1 bottom:1 taskar:1 fly:1 worst:1 went:1 substantial:1 dynamic:2 diversely:1 trained:5 depend:1 singh:1 segment:36 efficiency:1 learner:2 completely:1 easily:2 po:1 joint:1 represented:1 various:2 jersey:1 stacked:1 describe:1 artificial:3 labeling:2 outside:1 quite:2 encoded:1 loglikelihood:1 otherwise:2 statistic:1 itself:2 sequence:15 advantage:2 net:2 combining:1 pthe:1 intuitive:1 sourceforge:2 exploiting:2 ijcai:1 requirement:1 object:1 ac:1 nearest:1 sim:5 job:4 krogh:1 implemented:3 c:1 implies:2 capitalized:3 kraut:1 vx:1 human:1 f1:14 generalization:2 abbeel:1 tfidf:1 acapulco:1 extension:3 pl:1 strictly:1 hold:1 considered:4 viterbi:2 week:1 major:2 dictionary:26 early:1 estimation:1 proc:1 label:13 hs1:1 title:5 coordination:1 largest:1 city:7 always:2 gaussian:3 rather:3 probabilistically:1 focus:1 likelihood:2 indicates:2 contrast:1 normalizer:1 sigkdd:1 baseline:4 sense:1 helpful:1 inference:10 abstraction:1 mathematic:1 hidden:3 koller:1 quasi:1 overall:1 classification:2 uai02:1 integration:3 fairly:1 field:12 construct:3 never:1 extraction:11 identical:1 icml:3 future:1 np:2 sunita:2 individual:2 sterling:1 argmax:6 william:1 message:1 mining:1 evaluation:1 tj:7 hg:1 chain:6 partial:1 necessary:1 plotted:1 instance:2 formalism:1 column:2 bombay:1 markovian:7 cover:1 phrase:1 cost:3 stacking:1 organizational:1 subset:1 seventh:1 combined:1 person:7 espinosa:1 international:4 ie:1 probabilistic:3 together:1 precup:1 again:2 thesis:1 management:3 emnlp:1 external:10 creating:1 conf:1 li:1 potential:1 bfgs:1 student:1 int:7 satisfy:1 wcohen:1 depends:2 performed:2 start:1 hf:1 sort:2 iitb:1 contribution:1 formed:2 voted:4 accuracy:4 degraded:1 kaufmann:1 maximized:1 listing:1 conceptually:1 mooney:2 email:4 definition:1 sixth:2 against:1 associated:5 di:2 irvine:1 dataset:1 logical:1 recall:3 knowledge:2 segmentation:9 appears:1 higher:1 improved:1 formulation:1 evaluated:2 web:3 expressive:3 replacing:1 overlapping:1 name:11 omitting:1 usa:1 ranged:1 true:1 hence:2 assigned:2 excluded:2 conditionally:2 during:2 game:2 crf:38 performs:2 meaning:1 ranging:1 recently:1 common:2 cohen:3 tabulation:1 association:1 extend:2 analog:1 interpretation:1 numerically:1 kluwer:1 mellon:1 measurement:5 automatic:1 malouf:1 language:3 dj:3 stable:1 similarity:3 longer:1 base:5 segmental:1 closest:2 recent:1 skiing:1 certain:2 yi:4 morgan:1 additional:3 somewhat:1 preceding:1 maximize:3 fernando:1 july:1 semi:60 ii:1 faster:1 match:1 calculation:1 offer:1 academic:2 ravikumar:1 paired:1 variant:2 cmu:1 expectation:1 metric:4 grounded:1 source:1 appropriately:1 zw:5 rest:1 ascent:1 capitalization:3 december:1 lafferty:1 seem:1 call:1 extracting:1 near:1 easy:1 lerch:1 automated:1 xsj:5 architecture:1 opposite:1 inner:1 whether:1 nine:1 remark:1 generally:1 useful:1 clear:1 detailed:1 santa:1 bunescu:1 ten:2 reduced:2 http:2 outperform:3 percentage:1 diverse:1 carnegie:1 express:2 putting:1 four:3 traced:1 tenth:1 backward:2 nocedal:1 fraction:3 sum:4 run:1 package:2 letter:1 uncertainty:1 named:5 reasonable:1 home:1 decision:1 appendix:1 jaccard:1 conll:2 bound:1 replaces:1 inclusive:1 agichtein:1 tag:4 argument:1 span:3 concluding:1 performing:1 relatively:1 structured:2 marginbased:1 craven:1 smaller:1 y0:2 shallow:1 maxy:1 intuitively:1 restricted:1 pr:9 taken:1 chunking:2 fienberg:1 equation:4 letting:1 ge:1 tractable:1 end:2 available:8 rewritten:2 eight:1 hierarchical:2 appropriate:1 extern:2 ho:1 slower:1 denotes:1 include:3 linguistics:1 newton:1 exploit:1 sigmod:1 k1:1 uj:8 question:1 sha:2 primary:1 usual:2 gradient:2 deshmukh:1 distance:5 separate:1 simulated:1 entity:21 seven:1 mail:1 argue:1 trivial:1 toward:1 reason:1 induction:1 assuming:1 length:17 index:2 mexico:1 gk:1 implementation:3 collective:1 perform:3 allowing:1 upper:1 markov:25 datasets:3 extended:1 excluding:1 relational:4 team:1 canada:3 introduced:1 pair:2 sentence:1 california:1 elapsed:1 assembled:1 address:5 suggested:1 below:1 pattern:4 memory:2 power:3 dict:9 natural:7 indicator:7 representing:1 improve:2 technology:1 mdps:2 columbia:1 text:1 prior:2 discovery:2 relative:2 loss:1 interesting:1 proportional:1 triple:2 degree:1 s0:18 editor:1 elsewhere:1 placed:1 last:1 guide:1 allow:2 perceptron:6 institute:1 india:3 neighbor:3 curve:1 ending:3 valid:1 transition:2 genome:1 forward:2 made:3 reinforcement:2 sj:3 maxy0:1 gene:3 corpus:7 xi:7 discriminative:3 subsequence:3 table:5 learn:1 rohanimanesh:1 investigated:1 complex:1 sp:1 edition:1 grishman:1 edmonton:2 precision:2 position:5 pereira:4 wish:1 exponential:2 posting:1 british:1 specific:1 bishop:1 evidence:1 workshop:3 janssen:1 sequential:3 phd:1 entropy:2 wolpert:1 borkar:1 simply:1 expressed:1 contained:1 aa:1 corresponds:2 nested:2 extracted:1 acm:2 ma:1 conditional:12 goal:1 viewed:1 marked:2 change:2 specifically:2 uniformly:1 called:2 ewk:1 ew:3 internal:9 mark:1 brunswick:1 collins:1 indian:1 |
1,815 | 2,649 | Optimal information decoding from neuronal
populations with specific stimulus selectivity
Marcelo A. Montemurro
The University of Manchester
Faculty of Life Sciences
Moffat Building
PO Box 88, Manchester M60 1QD, UK
[email protected]
Stefano Panzeri ?
The University of Manchester
Faculty of Life Sciences
Moffat Building
PO Box 88, Manchester M60 1QD, UK
[email protected]
Abstract
A typical neuron in visual cortex receives most inputs from other cortical
neurons with a roughly similar stimulus preference. Does this arrangement of inputs allow efficient readout of sensory information by the target cortical neuron? We address this issue by using simple modelling of
neuronal population activity and information theoretic tools. We find that
efficient synaptic information transmission requires that the tuning curve
of the afferent neurons is approximately as wide as the spread of stimulus preferences of the afferent neurons reaching the target neuron. By
meta analysis of neurophysiological data we found that this is the case
for cortico-cortical inputs to neurons in visual cortex. We suggest that
the organization of V1 cortico-cortical synaptic inputs allows optimal information transmission.
1
Introduction
A typical neuron in visual cortex receives most of its inputs from other visual cortical neurons. The majority of cortico-cortical inputs arise from afferent cortical neurons with a
preference to stimuli which is similar to that of the target neuron [1, 2, 3]. For example, orientation selective neurons in superficial layers in ferret visual cortex receive more
than 50% of their cortico-cortical excitatory inputs from neurons with orientation preference which is less than 30o apart. However, this input structure is rather broad in terms
of stimulus-specificity: cortico-cortical connections between neurons tuned to dissimilar
stimulus orientation also exist [4]. The structure and spread of the stimulus specificity of
cortico-cortical connections has received a lot of attention because of its importance for
understanding the mechanisms of generation of orientation tuning (see [4] for a review).
However, little is still known on whether this structure of inputs allows efficient transmission of sensory information across cortico-cortical synapses.
It is likely that efficiency of information transmission across cortico-cortical synapses also
depends on the width of tuning curves of the afferent cortical neurons to stimuli. In fact,
theoretical work on population coding has shown that the width of the tuning curves has
?
Corresponding author
an important influence on the quality and the nature of the information encoding in cortical
populations [5, 6, 7, 8]. Another factor that may influence the efficiency of cortico-cortical
synaptic information transmission is the biophysical capability of the target neuron. To
conserve all information during synaptic transmission, the target neuron must conserve the
?label? of the spikes arriving from multiple input neurons at different places on its dendritic
tree [9]. Because of biophysical limitations, a target neuron that e.g. can only sum inputs
at the soma may lose a large part of the information present in the afferent activity. The
optimal arrangement of cortico-cortical synapses may also depend on the capability of
postsynaptic neurons in processing separately spikes from different neurons.
In this paper, we address the problem of whether cortico-cortical synaptic systems encode
information efficiently. We introduce a simple model of neuronal information processing
that takes into account both the selective distribution of stimulus preferences typical of
cortico-cortical connections and the potential biophysical limitations of cortical neurons.
We use this model and information theoretic tools to investigate whether there is an optimal trade-off between the spread of distribution of stimulus preference across the afferent
neurons and the tuning width of the afferent neurons itself. We find that efficient synaptic
information transmission requires that the tuning curve of the afferent neurons is approximately as wide as the spread of stimulus preferences of the afferent fibres reaching the
target neuron. By reviewing anatomical and physiological data, we argue that this optimal
trade-off is approximately reached in visual cortex. These results suggest that neurons in
visual cortex are wired to decode optimally information from a stimulus-specific distribution of synaptic inputs.
2
Model of the activity of the afferent neuronal population
We consider a simple model for the activity of the afferent neuronal population based on
the known tuning properties and spatial and synaptic organisation of sensory areas.
2.1
Stimulus tuning of individual afferent neurons
We assume that the the population is made of a large number N of neurons (for a real
cortical neuron, the number N of afferents is in the order of few thousands [10]). The
response of each neuron rk (k = 1, ? ? ? , N ) is quantified as the number of spikes fired in
a salient post-stimulus time window of a length ? . Thus, the overall neuronal population
response is represented as a spike count vector r = (r1 , ? ? ? , rN ).
We assume that the neurons are tuned to a small number D of relevant stimulus parameters
[11, 12], such as e.g. orientation, speed or direction of motion of a visual object. The
stimulus variable will thus be described as a vector s = (s1 , . . . , sD ) of dimension D. The
number of stimulus features that are encoded by the neuron will be left as a free parameter
to be varied within the range 1-5 for two reasons. First, although there is evidence that the
number of stimulus features encoded by a single neuron is limited [11, 12], more research
is still needed to determine exactly how many stimulus parameters are encoded in different
areas. Second, a previous related study [8] has shown that, when considering large neuronal
populations with a uniform distribution of stimulus preferences (such as an hypercolumn
in V1 containing all stimulus orientations) the tuning width of individual neurons which is
optimal for population coding depends crucially on the number of stimulus features being
encoded. Thus, it is interesting to investigate how the optimal arrangement of corticocortical synaptic systems depends on the number of stimulus features being encoded.
The neuronal tuning function of the k ? th neuron (k = 1, ? ? ? , N ), which quantifies the
mean spike count of the k ? th neuron to the presented stimulus, is modelled as a Gaussian
distribution, characterised by the following parameters: preferred stimulus s(k) , tuning
width ?f , and response modulation m:
?
f (k) (s) = me
(s?s(k) )2
2?f 2
(1)
The Gaussian tuning curve is a good description of the tuning properties of e.g. V1 or
MT neurons to variables such as stimulus orientation motion direction [13, 14, 15], and is
hence widely used in models of sensory coding [16, 17]. Large values of ?f indicate coarse
coding, whereas small values of ?f indicate sharp tuning.
Spike count responses of each neuron on each trial are assumed to follow a Poisson distribution whose mean is given by the above neuronal tuning function (Eq. 1). The Poisson
model is widely used because it is the simplest model of neuronal firing that captures the
salient property of neuronal firing that the variance of spike counts is proportional to its
mean. The Poisson model neglects all correlations between spikes. This assumption is
certainly a simplification but it is sufficient to account for the majority of the information
transmitted by real cortical neurons [18, 19, 20] and, as we shall see later, it is mathematically convenient because it makes our model tractable.
2.2
Distribution of stimulus preferences among the afferent population
Neurons in sensory cortex receive a large number of inputs from other neurons with a variety of stimulus preferences. However, the majority of their inputs come from neurons with
roughly similar stimulus preference [1, 2, 3]. To characterise correctly this type of spread
of stimulus preference among the afferent population, we assume (unlike in previous studies [8]), that the probability distribution of the preferred stimulus among afferent neurons
follows a Gaussian distribution:
P (?s) =
1
(2?)D/2 ?pD
?
e
(?
s??
s0 )2
2
2?p
(2)
In Eq. (2) the parameter ?s0 represents the the center of the distribution, thus being the
most represented preferred stimulus in the population. (we set, without loss of generality, ?s0 = 0.) The parameter ?p controls the spread of stimulus preferences of the afferent
neuronal population: a small value of ?p indicates that a large fraction of the population
have similar stimulus preferences, and a large value of ?p indicates that all stimuli are
represented similarly. A Gaussian distribution of stimulus preferences of the afferent population fits well empirical data on distribution of preferred orientations of synaptic inputs
of neurons in both deep and superficial layers of ferret primary visual cortex [3].
3
Width of tuning and spread of stimulus preferences in visual cortex
To estimate the width of tuning ?f and the spread of stimulus preferences ?p of corticocortical afferent populations in visual cortex, we reviewed critically published anatomical
and physiological data. We concentrated on excitatory synaptic inputs, which form the
majority of inputs to a cortical pyramidal neuron [10]. We computed ?p by fitting (by a least
square method) the published histograms of synaptic connections as function of stimulus
preference of the input neuron to Gaussian distributions. Similarly, we determined ?f by
fitting spike count histograms to Gaussian tuning curves.
When considering a target neuron in ferret primary visual cortex and using orientation as
the stimulus parameters, the spread of stimulus preferences ?p of its inputs is ? 20o for
layer 5/6 neurons [3], and 16o [3] to 23o [21] for layer 2/3 neurons. The orientation tuning
width ?f of the cortical inputs to the V1 target neuron is that of other V1 neurons that
project to it. This ?f is 17o for Layer 4 neurons [22], and it is similar for neurons in deep
and superficial layers [3]. When considering a target neuron in Layer 4 of cat visual cortex
and orientation tuning, the spread of stimulus preference ?p is 20o [2] and ?f is ? 17o .
When considering a target neuron in ferret visual cortex and motion direction tuning, the
spread of tuning of its inputs ?p is ? 30 o [1]. Motion direction tuning widths of macaque
neurons is ? 28o , and this width is similar across species (see [13]).
The most notable finding of our meta-analysis of published data is that ?p and ?f appear
to be approximately of the same size and their ratio ?f /?p is distributed around 1, in the
range 0.7 to 1.1 for the above data. We will use our model to understand whether this range
of ?f /?p corresponds to an optimal way to transmit information across a synaptic system.
4
Information theoretic quantification of population decoding
To characterise how a target neuronal system can decode the information about sensory
stimuli contained in the activity of its afferent neuronal population, we use mutual information [23]. The mutual information between a set of stimuli and the neuronal responses
quantifies how well any decoder can discriminate among stimuli by observing the neuronal
responses. This measure has the advantage of being independent of the decoding mechanism used, and thus puts precise constraints on the information that can be decoded by any
biological system operating on the afferent activity.
Previous studies on the information content of an afferent neuronal population [7, 8] have
assumed that the target neuronal decoding system can extract all the information during
synaptic transmission. To do so, the target neuron must conserve the ?label? of the spikes
arriving from multiple neurons at different sites on its dendritic tree [9]. Given the potential biophysical difficulty in processing each spike separately, a simple alternative to spike
labelling has been proposed, - spike pooling [10, 24]. In this scheme, the target neuron
simply sums up the afferent activity. To characterize how the decoding of afferent information would work in both cases, we compute both the information that can be decoded by
either a system that processes separately spikes from different neurons (the ?labeled-line
information?) and the information available to a decoder that sums all incoming spikes (the
?pooled information?) [9, 24]. In the next two subsections we define these quantities and
we explain how we compute it in our model.
4.1
The information available to the the labeled-line decoder
The mutual information between the set of the stimuli and the labeled-line neuronal population activity is defined as follows [9, 24]:
P (r|s)
LL
I (S, R) = dsP (s)
(3)
P (r|s) log
P (r)
r
where P (s) is the probability of stimulus occurrence (here taken for simplicity as a uniform distribution over the hypersphere of D dimensions and ?radius? s? ). P (r|s) is the
probability of observing a neuronal population response r conditional to the occurrence of
stimulus s, and P (r) = dsP (s)P (r|s). Since the response vector r keeps separate the
spike counts of each neuron, the amount of information in Eq. (3) is only accessible to a
decoder than can keep the label of which neuron fired which spike [9, 24]. The probability
P (r|s) is computed according to the Poisson distribution, which is entirely determined by
the knowledge of the tuning curves [5]. The labeled-line mutual information is difficult to
compute for large populations, because it requires the knowledge of the probability of the
large-dimensional response vector r. However, since in our model we assume that we have
a very large number of independent neurons in the population and that the total activity of
the system is of the order of its size, then we can use the following simpler (but still exact)
expression[16, 25]:
1
D
ln (2?e) +
ds P (s) ln (|J (s)|)
(4)
2
2
where H(S) is the entropy of the prior stimulus presentation distribution P (S), J (s) is
the Fisher information matrix and | . . . | stands for the determinant. The Fisher information
matrix is a D ? D matrix who?s elements i, j are defined as follows:
2
?
Ji,j (s) = ?
P (r|s)
log P (r|s) ,
(5)
?si sj
r
I LL (S, R) = H(S) ?
Fisher information is a useful measure of the accuracy with which a particular stimulus can
be reconstructed from a single trial observation of neuronal population activity. However,
in this paper it is used only as a step to obtain a computationally tractable expression for the
labeled-line mutual information. The Fisher information matrix can be computed by taking
into account that for a population of Poisson neurons is just the sum of the Fisher information for individual neurons, and the latter has a simple expression in terms of tuning curves
[16]. Since the neuronal population size N is is large, the sum over Fisher information
of individual neurons can be replaced by an integral over the stimulus preferences of the
neurons in the population, weighted by their probability density P (?s). After performing
the integral over the distribution of preferred stimuli, we arrived at the following result for
the elements of the Fisher information matrix:
?
? 2 1+?
N?m
? D?2
2)
2
(
?
+
?
(?
+
?
?
)
e
i,j
i,j
i
j
D
?p2 (1 + ? 2 ) 2 +2
2
Ji,j (s) =
(6)
where we have introduced the following short-hand notation ?f /?p ? ? and s/?p ? ?;
?i,j stands for the Kroneker Delta. From Eq. (6) it is possible to compute explicitly the
determinant |J (s)|, which has the following form:
|J (s)| =
D
?i = ?(?)D (1 + ? 2 )D?1 1 + ? 2 (1 + ? 2 )
(7)
i=1
where ?(?) is given by:
?
?
N?m
? D?2
e 2(1+?2 )
D
2
?p (1 + ? 2 ) 2 +1
2
?(?) =
(8)
Inserting Eq. (7) into Eq. (4), one obtains a tractable but still exact expression for the
mutual information , which has the advantage over Eq. (3) of requiring only an integral
over a D-dimensional stimulus rather than a sum over an infinite population.
We have studied numerically the dependence of the labeled-line information on the parameters ?f and ?p as a function of the number of encoded stimulus features D 1 . We
investigated this by fixing ?p and then varying the ration ?f /?p over a wide range. Results
(obtained for ?p = 1 but representative of a wide ?f range) are reported in Fig. 1. We
found that, unlike the case of a uniform distribution of stimulus preferences [8], there is a
finite value of the width of tuning ?f that maximizes the information for all D ? 2. Interestingly, for D ? 2 the range 0.7 ? ?f /?p ? 1.1 found in visual cortex either contains the
maximum or corresponds to near optimal values of information transmission. For D = 1,
information is maximal for very narrow tuning curves. However, also in this case the information values are still efficient in the cortical range ?f /?p ? 1, in that the tail of the
D = 1 information curve is avoided in that region. Thus, the range of values of ?f and ?p
found in visual cortex allows efficient synaptic information transmission over a wide range
of number of stimulus features encoded by the neuron.
We found (data not shown) that other parameters such as m and ? , had a weak or null effect on
the optimal configuration; see [17] for a D = 1 example in a different context.
1
ILL(S,R)
D=1
D=5
0
2
? /? 4
f
p
6
8
Figure 1: Mutual labeled-line information as a function of the ratio of tuning curve width
and stimulus preference spread ?f /?p . The curves for each stimulus dimensionality D
were shifted by a constant factor to separate them for visual inspection (lower curves correspond to higher values of D). The y-axis is thus in arbitrary units. The position of the
maximal information for each stimulus dimension falls either inside the range of values
of ?f /?p found in visual cortex, or very close to it (see text) . Parameters are as follows:
s? = 2, rmax = 50Hz, ? = 10ms.
4.2
The information available to the the pooling decoder
We now consider the case in which the target neuron cannot process separately spikes
from different neurons (for example, a neuron that just sums up post-synaptic potentials
of approximately equal weight at the soma). In this case the label of the neuron that fired
each spike is lost by the target neuron, and it can only operate on the pooled neuronal
signal, in which the identity of each spike is lost. Pooling mechanisms have been proposed
as simple information processing strategies for the nervous system. We now study how
pooling changes the requirements for efficient decoding by the target neuron.
Mathematically speaking,
pooling maps the vector r onto a scalar ? equal to the sum of the
individual activities: ? = rk . Thus, the mutual information that can be extracted by any
decoder that only pools it inputs is given by the following expression:
P (?|s)
(9)
I P (S, R) = dsP (s)
P (?|s) log
P (?)
?
where P (?|s) and P (?) are the the stimulus-conditional and stimulus-unconditional probability of observing a pooled population response ? on a single trial. The probability P (?|s)
can be computed by noting that a sum of Poisson-distributed responses is still a Poissondistributed response whose tuning curve to stimuli is just the sum of the individual tuning
curves. The pooled mutual information is thus a function of a single Poisson-distributed
response variables and can be computed easily also for large populations.
The dependence of the pooled information on the parameters ?f and ?p as a function of
the number of encoded stimulus features D is reported in Fig. 2. There is one important
difference with respect to the labeled-line information transmission case. The difference is
that, for the pooled information, there is a finite optimal value for information transmission
also when the neurons are tuned to one-dimensional stimulus feature. For all cases of stimulus dimensionality considered, the efficient information transmission though the pooled
IP(S,R)
D=1
D=3
0
1
2
?f/?p
3
4
Figure 2: Pooled mutual information as a function of the ratio of tuning curve width and
stimulus preference spread ?f /?p . The maxima are inside the range of experimental values
of ?f /?p found in the visual cortex, or very close to it (see text). As for Fig. 1, the curves
for each stimulus dimensionality D were shifted by a constant factor to separate them for
visual inspection (lower curves correspond to higher values of D). The y-axis is thus in
arbitrary units. Parameters are as follows: s? = 2, rmax = 50 Hz,? = 10ms.
neuronal decoder can still be reached in the visual cortical range 0.7 ? ?f ?p ? 1.1. This
finding shows that the range of values of ?f and ?p found in visual cortex allows efficient synaptic information transmission even if the target neuron does not rely on complex
dendritic processing to conserve the label of the neuron that fired the spike.
5
Conclusions
The stimulus specificity of cortico-cortical connections is important for understanding the
mechanisms of generation of orientation tuning (see [4]) for a review). Here, we have
shown that the stimulus-specific structure of cortico-cortical connections may have also implications for understanding cortico-cortical information transmission. Our results suggest
that, whatever the exact role of cortico-cortical synapses in generating orientation tuning,
their wiring allows efficient transmission of sensory information.
Acknowledgments
We thanks A. Nevado and R. Petersen for many interesting discussions. Research supported
by ICTP, HFSP, Royal Society and Wellcome Trust 066372/Z/01/Z.
References
[1] B. Roerig and J. P. Y. Kao. Organization of intracortical circuits in relation to direction
preference maps in ferret visual cortex. J. Neurosci., 19:RC44(105), 1999.
? T?oth, and Z. F. Kisv?arday. Orien[2] T. Yousef, T. Bonhoeffer, D-S. Kim, U. T. Eysel, E.
tation topography of layer 4 lateral networks revealed by optical imaging in cat visual
cortex (area 18). European J. Neurosci., 11:4291?4308, 1999.
[3] B. Roerig and B. Chen. Relations of local inhibitory and excitatory circuits to orientation preference maps in ferret visual cortex. Cerebral Cortex, 12:187?198, 2002.
[4] K. A. C. Martin. Microcircuits in visual cortex. Current Opinion in Neurobiology,
12:418?425, 2002.
[5] P. Dayan and L. F. Abbott. Theoretical Neuroscience. MIT Press, 2001.
[6] D. C. Fitzpatrick, R. Batra, T. R. Stanford, and S. Kuwada. A neuronal population
code for sound localization. Nature, 388:871?874, 1997.
[7] A. Pouget, S. Deneve, J-C. Ducom, and P.E. Latham. Narrow versus wide tuning
curves: what?s best for a population code? Neural Computation, 11:85?90, 1999.
[8] K. Zhang and T.J. Sejnowski. Neuronal tuning: to sharpen or to broaden? Neural
Computation, 11:75?84, 1999.
[9] D. S. Reich, F. Mechler, and J. D. Victor. Independent and redundant information in
nearby cortical neurons. Science, 294:2566?2568, 2001.
[10] M. N. Shadlen and W. T. Newsome. The variable discharge of cortical neurons:
implications for connectivity, computation and coding. J. Neurosci., 18(10):3870?
3896, 1998.
[11] N. Brenner, W. Bialek, and R. de Ruyter van Steveninck. Adaptive rescaling maximizes information transmission. Neuron, 26:695?702, 2000.
[12] J. Touryan, B. Lau, and Y. Dan. Isolation of relevant visual features from random
stimuli for cortical complex cells. J. Neurosci, 22:10811?10818, 2002.
[13] T. D. Albright. Direction and orientation selectivity of neurons in visual area MT of
the macaque. J. Neurophysiol., 52:1106?1130, 1984.
[14] K.H. Britten, M. N. Shadlen, W. T. Newsome, and J. A. Movshon. The analysis
of visual-motion - a comparison of neuronal and psychophysical performance. J.
Neurosci., 12:4745?4765, 1992.
[15] K Kang, RM Shapley, and H Sompolinsky. Information tuning of population of neurons in primary visual cortex. J. Neurosci., 24:3726?3735, 2004.
[16] N. Brunel and J. P. Nadal. Mutual information, fisher information and population
coding. Neural Computation, 10:1731?1757, 1998.
[17] A. Nevado, M.P. Young, and S. Panzeri. Functional imaging and neural information
coding. Neuroimage, 21:1095?1095, 2004.
[18] S. Nirenberg, S. M. Carcieri, A.L. Jacobs, and P. E. Latham. Retinal ganglion cells
act largely as independent encoders. Nature, 411:698?701, 2001.
[19] R. S. Petersen, S. Panzeri, and M.E. Diamond. Population coding of stimulus location
in rat somatosensory cortex. Neuron, 32:503?514, 2001.
[20] M. W. Oram, N.G. Hatsopoulos, B.J. Richmond, and J.P. Donoghue. Excess synchrony in motor cortical neurons provides redundant direction information with that
from coarse temporal measures. J. Neurophysiol., 86:1700?1716, 2001.
[21] M. B. Dalva, M. Weliky, and L. Katz. Relations between local synaptic connections
and orientation domains in primary visual cortex. Neuron, 19:871?880, 1997.
[22] W. M. Usrey, M. P. Sceniak, and B. Chapman. Receptive fields and response properties of neurons in layer 4 of ferret visual cortex. J. Neurophysiol., 89:1003?1015,
2003.
[23] T.M. Cover and J.A. Thomas. Elements of information theory. John Wiley, 1991.
[24] S. Panzeri, F. Petroni, R.S. Petersen, and M.E. Diamond. Decoding neuronal population activity in rat somatosensory cortex: role of columnar organization. Cerebral
Cortex, 13:45?52, 2003.
[25] K Kang and H Sompolinsky. Mutual information of population codes and distance
measures in probability space. Phys. Rev. Lett., 86:4958?4961, 2001.
| 2649 |@word trial:3 determinant:2 faculty:2 crucially:1 jacob:1 configuration:1 contains:1 tuned:3 interestingly:1 current:1 si:1 must:2 john:1 motor:1 nervous:1 inspection:2 short:1 hypersphere:1 provides:1 coarse:2 location:1 preference:26 simpler:1 zhang:1 fitting:2 dan:1 shapley:1 inside:2 introduce:1 montemurro:2 roughly:2 little:1 window:1 considering:4 project:1 notation:1 maximizes:2 circuit:2 null:1 what:1 rmax:2 nadal:1 finding:2 temporal:1 m60:2 act:1 exactly:1 rm:1 uk:4 control:1 unit:2 whatever:1 appear:1 local:2 sd:1 tation:1 encoding:1 firing:2 modulation:1 approximately:5 studied:1 quantified:1 limited:1 range:13 steveninck:1 acknowledgment:1 lost:2 area:4 empirical:1 convenient:1 specificity:3 carcieri:1 suggest:3 petersen:3 cannot:1 close:2 onto:1 put:1 context:1 influence:2 map:3 center:1 attention:1 simplicity:1 pouget:1 population:38 transmit:1 discharge:1 target:19 decode:2 exact:3 element:3 conserve:4 corticocortical:2 labeled:8 role:2 capture:1 thousand:1 readout:1 region:1 sompolinsky:2 trade:2 hatsopoulos:1 pd:1 ration:1 depend:1 reviewing:1 localization:1 efficiency:2 neurophysiol:3 po:2 easily:1 represented:3 cat:2 sejnowski:1 whose:2 encoded:8 widely:2 stanford:1 nirenberg:1 itself:1 ip:1 advantage:2 biophysical:4 maximal:2 inserting:1 relevant:2 eysel:1 fired:4 description:1 kao:1 manchester:6 transmission:17 r1:1 requirement:1 wired:1 generating:1 object:1 ac:2 fixing:1 received:1 eq:7 p2:1 indicate:2 come:1 qd:2 somatosensory:2 direction:7 radius:1 opinion:1 dendritic:3 biological:1 mathematically:2 around:1 considered:1 panzeri:5 fitzpatrick:1 lose:1 label:5 tool:2 weighted:1 mit:1 gaussian:6 reaching:2 rather:2 varying:1 encode:1 dsp:3 modelling:1 indicates:2 richmond:1 kim:1 dayan:1 poissondistributed:1 relation:3 selective:2 issue:1 overall:1 orientation:16 among:4 ill:1 spatial:1 mutual:12 equal:2 field:1 chapman:1 represents:1 broad:1 stimulus:71 few:1 individual:6 replaced:1 organization:3 investigate:2 certainly:1 unconditional:1 oth:1 implication:2 integral:3 moffat:2 tree:2 theoretical:2 cover:1 newsome:2 uniform:3 optimally:1 characterize:1 reported:2 encoders:1 thanks:1 density:1 accessible:1 off:2 decoding:7 pool:1 connectivity:1 containing:1 rescaling:1 account:3 potential:3 de:1 intracortical:1 retinal:1 coding:8 pooled:8 notable:1 explicitly:1 afferent:24 depends:3 later:1 lot:1 observing:3 reached:2 capability:2 synchrony:1 marcelo:1 square:1 accuracy:1 variance:1 who:1 efficiently:1 largely:1 correspond:2 modelled:1 weak:1 critically:1 published:3 explain:1 synapsis:4 phys:1 synaptic:18 oram:1 subsection:1 knowledge:2 dimensionality:3 higher:2 follow:1 response:14 box:2 though:1 generality:1 microcircuit:1 just:3 correlation:1 d:1 hand:1 receives:2 trust:1 quality:1 building:2 effect:1 requiring:1 hence:1 wiring:1 ll:2 during:2 width:13 rat:2 m:2 arrived:1 theoretic:3 latham:2 motion:5 stefano:1 functional:1 mt:2 ji:2 cerebral:2 tail:1 katz:1 numerically:1 tuning:36 similarly:2 sharpen:1 had:1 reich:1 cortex:29 operating:1 apart:1 selectivity:2 meta:2 sceniak:1 life:2 victor:1 transmitted:1 determine:1 redundant:2 signal:1 multiple:2 sound:1 post:2 poisson:7 histogram:2 cell:2 receive:2 whereas:1 separately:4 ferret:7 pyramidal:1 touryan:1 operate:1 unlike:2 pooling:5 hz:2 near:1 noting:1 revealed:1 variety:1 fit:1 isolation:1 donoghue:1 whether:4 expression:5 movshon:1 speaking:1 deep:2 useful:1 characterise:2 amount:1 concentrated:1 simplest:1 exist:1 inhibitory:1 shifted:2 delta:1 neuroscience:1 correctly:1 anatomical:2 shall:1 salient:2 soma:2 abbott:1 v1:5 imaging:2 deneve:1 fraction:1 sum:10 fibre:1 place:1 entirely:1 layer:9 simplification:1 activity:12 constraint:1 petroni:1 nearby:1 speed:1 performing:1 optical:1 martin:1 according:1 mechler:1 across:5 postsynaptic:1 rev:1 s1:1 lau:1 taken:1 wellcome:1 ln:2 computationally:1 count:6 mechanism:4 needed:1 tractable:3 available:3 occurrence:2 alternative:1 thomas:1 broaden:1 neglect:1 society:1 psychophysical:1 arrangement:3 quantity:1 spike:21 strategy:1 primary:4 dependence:2 receptive:1 bialek:1 distance:1 separate:3 lateral:1 majority:4 decoder:7 me:1 argue:1 reason:1 length:1 code:3 ratio:3 difficult:1 yousef:1 diamond:2 neuron:87 observation:1 finite:2 neurobiology:1 precise:1 rn:1 varied:1 sharp:1 arbitrary:2 introduced:1 hypercolumn:1 connection:7 kroneker:1 narrow:2 kang:2 macaque:2 address:2 royal:1 difficulty:1 quantification:1 rely:1 ducom:1 scheme:1 axis:2 britten:1 extract:1 text:2 review:2 understanding:3 prior:1 loss:1 topography:1 generation:2 limitation:2 interesting:2 proportional:1 versus:1 sufficient:1 weliky:1 s0:3 shadlen:2 excitatory:3 supported:1 free:1 arriving:2 allow:1 cortico:16 understand:1 wide:6 fall:1 taking:1 distributed:3 van:1 curve:19 dimension:3 cortical:33 stand:2 lett:1 sensory:7 author:1 made:1 adaptive:1 avoided:1 sj:1 reconstructed:1 obtains:1 excess:1 preferred:5 keep:2 incoming:1 assumed:2 quantifies:2 reviewed:1 nature:3 superficial:3 ruyter:1 investigated:1 complex:2 european:1 domain:1 spread:13 neurosci:6 arise:1 usrey:1 neuronal:28 site:1 representative:1 fig:3 wiley:1 neuroimage:1 position:1 decoded:2 young:1 rk:2 specific:3 physiological:2 organisation:1 evidence:1 importance:1 labelling:1 chen:1 columnar:1 entropy:1 bonhoeffer:1 simply:1 likely:1 ganglion:1 neurophysiological:1 visual:32 contained:1 scalar:1 brunel:1 corresponds:2 extracted:1 conditional:2 identity:1 presentation:1 fisher:8 content:1 change:1 brenner:1 typical:3 characterised:1 determined:2 infinite:1 total:1 specie:1 discriminate:1 hfsp:1 experimental:1 batra:1 albright:1 latter:1 dissimilar:1 |
1,816 | 265 | 542
Kassebaum, Thnorio and Schaefers
The Cocktail Party Problem:
Speech/Data Signal Separation Comparison
between Backpropagation and SONN
John Kassebaum
[email protected]
Manoel Fernando Tenorio
[email protected]
Christoph Schaefers
Parallel Distributed Structures Laboratory
School of Electrical Engineering
Purdue University
W. Lafayette, IN. 47907
ABSTRACT
This work introduces a new method called Self Organizing Neural
Network (SONN) algorithm and compares its performance with Back
Propagation in a signal separation application. The problem is to
separate two signals; a modem data signal and a male speech signal,
added and transmitted through a 4 khz channel. The signals are sampled at 8 khz, and using supervised learning, an attempt is made to
reconstruct them. The SONN is an algorithm that constructs its own
network topology during training, which is shown to be much smaller
than the BP network, faster to trained, and free from the trial-anderror network design that characterize BP.
1. INTRODUCTION
The research in Neural Networks has witnessed major changes in algorithm design
focus, motivated by the limitations perceived in the algorithms available at the
time. With the extensive work performed in that last few years using multilayered
networks, it was soon discovered that these networks present limitations in tasks
The Cocktail Party Problem:
that: (a) are difficult to determine problem complexity a priori, and thus design
network of the correct size, (b) training not only takes prohibitively long times,
but requires a large number of samples as well as fine parameter adjustment,
without guarantee of convergence, (c) such networks do not handle the system
identification task efficiently for systems whose time varying structure changes
radically, and, (d) the trained network is little more than a black box of weights
and connections, revealing little about the problem structure; being hard to find
the justification for the algorithm weight choice, or an explanation for the output
decisions based on an input vector. We believe that this need is sparking the
emergence of a third generation of algorithms to address such questions.
2. THE SELF
ALGORITHM
ORGANIZING
NEURAL
NETWORK
2.1 SELF ORGANIZING NETWORK FAMILY
A family of Self Organizing Structure (SOS) Algorithms can be readily designed
with our present knowledge, and can be used as a tool to research the motivating
questions. Each individual algorithm in this family might have different characteristics, which are summarized in the following list:
- A search strategy for the structure of the final model
- A rule of connectivity
- A performance criteria
- A transfer function set with appropriate training rule
As we will show here, by varying each one of these components, a different
behavior of the algorithm can be imposed.
Self organizing structure algorithms are not new. These algorithms have been
present in the statistical literature since the mid 70's in a very different context.
As far as we know, the first one to propose such an algorithm was Ivahnenko
[1971] which was followed by a host of variations on that original proposal
[Duffy&Franklin, 1975; Ikeda, et al., 1976; Tomura&Kondo, 1980; Farlow,1989].
Ivahnenko's subfamily of algorithms (GMDH - Group Method of Data Handling)
can be characterized in our classification by the same four-tuple criterion: (1) gradient descent local search, (2) creation of regular feedforward layers with elements
pairwisely connected, (3) least-mean-squares estimation, and (4) a single element
set comprised of a 2 order bivariate function.
Here we want to present our subfamily (SON - Self Organizing Networks) of the
SOS algorithm family, characterized differently by: (1) global optimization search,
(2) arbitrary connectivity based on an arbitrary number of neuron inputs, (3)
Structure Estimation Criteria (SEC) (a variation of Rissanen's [1983]. Minimum
Description Length Criteria, extended to the hierarchical case), and, (4) for training speed, activation functions are restricted to be linear on the parameters and
the output functions need to be invertible, no other restriction is imposed in kind
or number. The particular algorithm presented here is called the Self Organizing
543
544
Kassebaum, Tenorio and Schaefers
Neural Network (SONN) [Tenorio&Lee, 1988,1989; Tenorio 1990 a,b]. It was composed of: (1) a graph synthesis procedure based on Simulated Annealing [Kirkpatrick et a.1., 1983]; (2) two input neurons that a.re arbitrarily connected; (3) the
Structure Estimation Criteria; a.nd, (4) a set of a.ll polynomials that a.re special
ca.ses of 2nd order bivariates a.nd inclusive, followed or not by sigmoid functions.
The SONN a.lgorithm performs a. search in the model space by the construction
of hypersurfa.ces. A network of nodes, each node representing a. hypersurface, is
organized to be a.n a.pproximate model of the real system. Below, the components
of SONN a.re discussed.
2.2 THE ALGORITHM STRUCTURE
The mechanisms behind the a.lgorithm works as follows. First, create a. set of terminals which a.re the output of the nodes a.vailable for connection to other nodes.
This set is initialized with the output of the input nodes; in other words, the input
variables themselves. From this set, with uniform probability, select a subset (2 in
our case) of terminals, a.nd used them as inputs to the new node. To construct the
new node, select a.ll the function of the set of prototype functions (activation followed by output function), a.nd evaluate the SEC using the terminals as inputs.
Selecting the best function, test for the acceptance of that node according to the
Simulated Annealing move a.cceptance criterion. If the new node is a.ccepted, place
its output in the set of terminals and iterate until the optimum model is found.
The details or the a.lgorithm can be found in [Tenorio&Lee, 1989].
2.2.1 The Prototype Functions
Consider the Mahalanobis distance:
Yj
=sig{(x-/-LPC- 1 (x-/-L)t}
(1 )
This distance ca.n be rewritten as a second order function, whose parameters are
the indirect representation of the covariance matrix X and the mean vector /-L.
This function is linear in the parameters, which makes it easy to perform training,
a.nd it is the function with the smallest degree of non linearity; only simpler is the
linear case. Interestingly enough, this is the same prototype function used in the
GMDH a.lgorithm to form the Ivahnenko polynomial for apparently completely
different reasons. In the SONN, this function is taken to be 2-input and all its possible variations (32) by setting parameters to zero are included in the set of
a.ctivation functions. This set combined with the output function (the identify or
sigmoid), for the set of prototype functions, used by the a.lgorithm in the node
construction.
2.2.2 Evaluation of the Model Based on the MDL Criterion
The selection rule of the neuron transfer function was based on a modification of
the Minimal Description Length (MOL) information criterion. In [Rissanen, 1978],
the principle of minimal description for statistical estimation was developed. The
reason for the choice of such a criterion is that, in general the accuracy of the
model can increase at the expense of simplicity in the number oC parameters. The
The Cocktail Party Problem:
increase of complexity might also be accompanied by the overfitting of the model.
To overcome this problem, the MDL provides a trade-oft' between the accuracy
and the complexity of the model by including the structure estimation term of the
final model. The final model (with the minimal MDL) is optimum in the sense of
being a consistent estimate of the number of parameters while achieving the
minImUm error [Rissanen, 1980]. Given a sequence of observations
Xl ,X2 , ??? ,XN from the random variable X, the dominant term of the MDL in
[Rissanen, 1978] is:
MDL = -log f(x
Ie) +0.5 k log N
(2)
Ie)
where f(x
is the estimated probability density function of the model, k is the
number of parameters, and N is the number of observations. The first term is
actually the negative of the maximum likelihood (ML) with respect to the
estimated parameter. The second term describes the structure of the models and it
is used as a penalty for the complexity of the model.
3. EXAMPLE - THE COCKTAIL PARTY PROBLEM
The Cocktail Party Problem is the name given to the phenomenon that people
can understand and track speech in a noisy environment, even when the noise is
being made by other speakers. A simpler version of this problem is presented here:
a 4 khz channel is excited with male speech and modem data additively at the
same time. The task presented to the network is to separate both signals.
To compare the accuracy of the signal separation between the SONN and the
Back Propagation algorithms a normalized RMSE is used as a performance
index:
normalized RMSE _ _ _ _R_M_S_E
_ __
StandardDevision
(3)
3.1. EXPERIMENTS WITH BACK PROPAGATION
In order to design a filter using Back Propagation for this task, several architectures were considered. Since the input and output to the problem are time series,
and such architectures are static, modifications to the original paradigm is
required to deal with the time dimension. Several proposals have been made in
this respect: tapped delay filters, recurrent architectures, low pass filter transfer
functions, modified discriminant functions, and self excitatory connections (see
[Wah, Tenorio, Merha, and Fortes, 90] ). The best result for this task was
achieved by two tapped delay lines in the input layer, one for the input signal, the
other for the output signal. The network was trained to recognize the speech signal from the mixed signal. The mixed signal had a speech to modem data energy
ratio of 4:1, or 2.5 dB.
The network was designed to be a feedforward with 42 inputs (21 delayed versions
of the input signal, and similarly for the output signal), 15 hidden units, and a
single output unit. The network was trained with a single phoneme, taking about
545
546
Kassebaum, Tenorio and Schaefers
10 cpu-hours on a Sequent machine. The network when presented with the trained
phoneme added to the modem data, produced a speech reconstruct ability error
equal to a nRMSE of 0.910. Previously several different configurations of the network were tried as well as different network parameters, and signal ratios of 1:1;
all with poor results. Few networks actually converged to a final solution. A
major problem with the BP architecture is that it can perfectly filter the signal in
the first few samples, just to later demonstrate increasing amounts of cumulative
errors; this instability may be fruit of the recurring nature of the architecture,
and suboptimal weight training (Figure 2). The difficulty in finding and fine tuning the architecture, the training convergence, and time requirements led us to
later stop pursuing the design of these filters with Back Propagation strategies.
3.2. EXPERIMENTS WITH SONN
At that time, the SONN algorithm had been successfully used for identification
and prediction tasks [Tenorio&Lee; 88,89,90]. To make the task more realistic
with possible practical utilization of this filter (Data-Over-Voice Circuits), the
energy ratio between the voice and the modem data was reduced to 1:1, or 0 dB.
A tapped delay line containing 21 delayed versions of the mixed signal was
presented to the algorithm. Two sets of prototype functions were used, and both
contained the full set of 32 variations of 2nd order bivariates. The first set had
the identity (SONN-I experiments) and the second had a sigmoid (SONN-SIG
experiments) as the output function for each node.
SONN-I created 370 nodes, designing a final model with 5 nodes. The final symbolic transfer function which represents the closed form function of the network
was extracted. Using a Gould Powernode 9080, this search took 98.6 sec, with an
average of 3.75 nodes/sec. The final model had an nRMSE of 0.762 (Figure 3) for
reconstructed speech with the same BP data; with 19 weights. Training with the
modem signalled to nRMSE of 0.762 (Figure 4) for the BP data. A search using
the SONN-SIG model was allowed to generate 1000 nodes, designing a final model
with 5 nodes. With the same computer, the second search took 283.42 sec, with an
average 3.5 nodes/sec. The final model had an nRMSE comparable to the SONN-I
(better by 5-10%); with 20 weights. The main characteristics of both signals were
captured, specially if one looks at the plots and notices the same order of nonlinearity between the real and estimated signals (no over or under estimation).
Because of the forgiving nature of the human speech perception, the voice after
reconstruction, although sightly muffled, remains of good quality; and the reconstructed modem signal can be used to reconstruct the original digital message,
without much further post processing. The SONN does not present cumulative
errors during the reconstruction, and when test with different (unseen, from the
same speaker) speech data, performed as well as with the test data. We have yet
to fully explore the implication of that to different speakers and with speaker of
different gender or language. These results will be reported elsewhere.
4. COMPARISON BETWEEN THE TWO ALGORITHMS
Below we outline the comparison between the two algorithms drawn from our
experience with this signal separation problem.
The Cocktail Party Problem:
4.1. ADVANTAGES
The following were advantages of the SONN approach over the BP paradigm.
The most striking difference was found in the training times, and in the amount
of data required for training. The BP required 42 inputs (memories), where as the
SONN functioned with 21 inputs, actually using as few as 4 in the final model
(input variable selection). The SONN removed the problem of model estimation
and architecture design. The number of connections with the SONN models is as
low as 8 for 20 weights (relevant connections), as compared with 645 connections
and weights for the BP model. The accuracy and complexity of the model can be
trade for learning time as in BP, but the models that were more accurate also
required less parameters than BP. The networks are not required to be homogeneous, thus contributing to smaller models as well. Above all, the SONN can produce both the C code for the network as well as the sequence of individual node
symbolic functions; the SONN-I can also produce the symbolic representation of
the closed form function of he entire network.
4.2. DISADVANTAGES
Certain disadvantages of using self-organizing topology networks with stochastic
optimization algorithms were also apparent. The learning time of the SONN is
non deterministic, and depends on the model complexity and starting point.
Those are characteristic of the Simulated Annealing (SA) algorithm. These disadvantages are also present in the BP approach for different reasons. The connectivity of the model is not known a priori, which does not permit hardware implementation algorithms with direct connectivity emulation. Because the SONN
selects nodes from a growing set with uniform probability, the probability of
choosing a pair of nodes decreases with the inverse of the square of the number of
nodes. Thus algorithm effectiveness decreases with processing time. Careful plotting of the SEC, nRMSE, and complexity trajectories during training reveal that
the first 10% of the processing time achieves 90% of the final steady state values.
Biasing the node selection procedure might be an alternative to modify this
behavior. Simulated Annealing also required parametric tuning of the algorithm
by setting" the initial and final temperature, the duration of the search at each
temperature and the temperature decay. Alternative algorithms such as A * might
produce a better alternative to stochastic search algorithms.
6. CONCLUSION AND FUTURE WORK
In this study, we proposed a new approach for the signal separation filter design
neural network (SONN) algorithm. The varibased on a flexible, self-organizi
able structure provides the oppo; llity to search and construct the optimal model
based on input-output observations. The hierarchical v' ton of the MDL, ' lIed
the Structure Estimation Criteria, was used to guide
.; trade-off betwel the
model complexity and the accuracy of the estimation. The SONN approach
demonstrates potential usefulness as a tool for non linear signal processing function design.
We would like to explore the use of high level knowledge for function selection
547
548
Kassebaum, Thnorio and Schaefers
and connectivity. Also, the issues involving estimator and deterministic searches
are still open. Currently we are exploring the use of SONN for digital circuit synthesis, and studying how close the architecture generated here can approach the
design of natural structures when performing similar functions. More classification
problems, and problems involving dynamical systems (adaptive control and signal
processing) need to be explored to give us the experience needed to tackle the
problems for which it was designed.
6.
NOTE
The results reported here were originally intended for two papers accepted for
presentation at the NIPS'89. The organizing committee asked us to fuse the into a
single presentation for organizational purposes. In the limited time and the small
space allocated for the presentation of these results, we sought a compromise
between the reporting of the results and the description and comments on our
experience with the algorithm. The interested reader should look at the other
references about the SONN listed here and forthcoming papers.
REFERENCE
A. G. Ivakhnenko, (1971) "Polynomial Theory of Complex Systems," IEEE Trans.
S.M.C, Vol. SMC-1, no.4, pp. 364-378, Oct.
J. J. Duffy and M. A. Franklin, (1975) "A Learning Identification Algorithm and
its Application to an Environmental System," IEEE Trans. S. M. C., Vol. SMC-5,
no. 2, pp. 226-240.
S. Ikeda, M. Ochiai and Y. Sawarogi, (1976) "Sequential GMDH Algorithm and its
Application to River Flow Prediction," IEEE Trans S.M.C., Vol. SMC-6, no.7, pp.
473-479, July.
H. Tamura, T. Kondo, (1980) "Heuristics Free Group Method of Data Handling
Algorithm of Generating Optimal Partial Polynomials with Application to Air
Pollution Predication," Int. J. Systems Sci., 11,no.9, pp. 1095-1111.
J. Rissanen (1978) "Modeling by Shortest Data Description," Automatica, Vol.14,
pp. 465-471.
J. Rissanen, (1980) "Consistent Order Estimation of Autoregression Processes by
Shortest Description of Data," Analysis and Optimation of Stochastic System,
Jacobs et al eds. NY Academic.
J. Rissanen, (1983) "A Universal Prior for Integers and Estimation by Minimum
Description Length," Annuals of Statistics, Vol. 11 , no. 2, pp.416-431.
S.Kirkpatrick, C.D. Gelatt, M.P. Vecchi, (1983) "Optimization by Simulated
Annealing," Science, vol.220, pp. 671-680, May.
M. F. M. Tenorio and W.-T. Lee, (1988) "Self-Organizing Neural Network for the
Identification Problem," Advances in Neural Information Processing Systems I,
David S. Touretzky ed., pp. 57-64.
M. F. M. Tenorio and W.-T. Lee, (1989) "Self-Organizing Neural Network for the
The Cocktail Party Problem:
549
Identification Problem," School of Electrical Engineering, Purdue University,
Tech Report TR-EE 89-20, June.
M. F. M. Tenorio and W. -T. Lee, (1990) "Self-Organizing Network for the
Identification Problem," (expanded) IEEE Trans. on the Neural Networks, to
appear.
M. F. M. Tenorio, (1990) "The Self-Organizing Neural Network Algorithm:
Adapting Topology for Optimum Supervised Learning," IEEE Hawaii Conference
in Systems Science, 22, January.
M. F. Tenorio, (1990) "Self-Organizing Neural Network for the Signal Separation
Problem," to be submitted.
B. Wah, M. Tenorio, P. Mehra, J. Fortes, (1990) "Artificial Neural Networks:
Theory, Algorithms, Application and Implementations," IEEE press.
-----0..._
....u?el
A Il ??ill
A 19 ?? (l-S)
.2laJl(l-61
1110
axi+bX4X'!2+CX4X t 3+dX4X241'eX4X t 7+t'x.+lXf3+hxI3Xl4+ixI3
Q+-______----__----__ o
+jxU+kx I'7+m
Fl.- I: The SONNoSlO Nawark
FOnD
aaG . .
SO... T,.... FraM ..... ~ to""
?
zoo
lOa
T-._
s.-cn
0... S - 0 . . . 8,. AIoOnI/IIII
JIlt
i..
,.a
~
SONN?I SynIOOIic CloIed
_
-
SONN T,.... FraM ..... SlQNlto .....
D.JIlt
. ?s....O. . . . "'go iI...
_
-0..._
----
c
I
'r
I
- - _ _ 0-
'I
'I
!2111
~
-
..!
a
__
----~O
'.0T-. _ _zao
l ..
'i
I
~
t
.5' 1110
+-----~----
a
a
;
101
.
.
- - 0..._0-
"
1110
Q+-----~---~
' .0
__----__
zoo
T-._
O
lao
i
| 265 |@word trial:1 version:3 polynomial:4 nd:7 open:1 additively:1 tried:1 covariance:1 excited:1 jacob:1 tr:1 initial:1 configuration:1 series:1 selecting:1 interestingly:1 franklin:2 kondo:2 activation:2 yet:1 readily:1 john:1 ikeda:2 realistic:1 fram:2 designed:3 plot:1 provides:2 node:22 simpler:2 direct:1 behavior:2 themselves:1 growing:1 terminal:4 little:2 cpu:1 increasing:1 linearity:1 circuit:2 kind:1 developed:1 finding:1 guarantee:1 tackle:1 prohibitively:1 demonstrates:1 utilization:1 unit:2 control:1 appear:1 engineering:2 local:1 modify:1 farlow:1 black:1 might:4 christoph:1 limited:1 smc:3 lafayette:1 practical:1 yj:1 backpropagation:1 procedure:2 universal:1 adapting:1 revealing:1 word:1 regular:1 symbolic:3 close:1 selection:4 context:1 instability:1 restriction:1 imposed:2 deterministic:2 lied:1 go:1 starting:1 duration:1 simplicity:1 rule:3 estimator:1 handle:1 variation:4 justification:1 construction:2 homogeneous:1 designing:2 sig:3 tapped:3 element:2 electrical:2 connected:2 trade:3 removed:1 decrease:2 environment:1 complexity:8 asked:1 trained:5 compromise:1 creation:1 completely:1 indirect:1 differently:1 artificial:1 choosing:1 whose:2 apparent:1 heuristic:1 s:1 reconstruct:3 ability:1 statistic:1 unseen:1 emergence:1 noisy:1 final:12 sequence:2 advantage:2 took:2 propose:1 reconstruction:2 relevant:1 organizing:14 anderror:1 description:7 gmdh:3 convergence:2 optimum:3 requirement:1 produce:3 generating:1 recurrent:1 school:2 sa:1 sonn:30 emulation:1 correct:1 filter:7 stochastic:3 human:1 sightly:1 fond:1 exploring:1 considered:1 major:2 achieves:1 sought:1 smallest:1 purpose:1 perceived:1 estimation:11 currently:1 create:1 successfully:1 tool:2 modified:1 varying:2 focus:1 june:1 likelihood:1 tech:1 sense:1 fortes:2 el:1 entire:1 hidden:1 selects:1 interested:1 issue:1 classification:2 flexible:1 ill:1 priori:2 special:1 equal:1 construct:3 represents:1 look:2 future:1 report:1 few:4 composed:1 recognize:1 individual:2 delayed:2 pproximate:1 intended:1 attempt:1 acceptance:1 message:1 evaluation:1 mdl:6 introduces:1 male:2 kirkpatrick:2 signalled:1 behind:1 implication:1 accurate:1 tuple:1 partial:1 experience:3 initialized:1 re:4 minimal:3 witnessed:1 modeling:1 disadvantage:3 organizational:1 subset:1 uniform:2 comprised:1 delay:3 usefulness:1 characterize:1 motivating:1 reported:2 combined:1 density:1 river:1 ie:2 lee:6 off:1 invertible:1 synthesis:2 connectivity:5 containing:1 hawaii:1 potential:1 accompanied:1 summarized:1 sec:7 int:1 depends:1 performed:2 later:2 closed:2 apparently:1 parallel:1 kassebaum:5 rmse:2 square:2 air:1 accuracy:5 il:1 phoneme:2 characteristic:3 efficiently:1 identify:1 identification:6 produced:1 zoo:2 trajectory:1 converged:1 submitted:1 touretzky:1 ed:2 energy:2 pp:8 static:1 sampled:1 stop:1 knowledge:2 organized:1 actually:3 back:5 originally:1 supervised:2 box:1 just:1 until:1 propagation:5 quality:1 reveal:1 believe:1 name:1 normalized:2 lgorithm:5 laboratory:1 deal:1 mahalanobis:1 ll:2 during:3 self:15 speaker:4 steady:1 oc:1 criterion:10 outline:1 demonstrate:1 performs:1 temperature:3 aag:1 sigmoid:3 ecn:2 khz:3 discussed:1 he:1 tuning:2 similarly:1 nonlinearity:1 language:1 had:6 dominant:1 own:1 certain:1 arbitrarily:1 transmitted:1 minimum:3 captured:1 determine:1 shortest:2 fernando:1 paradigm:2 july:1 ii:1 signal:26 full:1 faster:1 characterized:2 academic:1 long:1 host:1 post:1 prediction:2 involving:2 achieved:1 tamura:1 proposal:2 want:1 fine:2 iiii:1 annealing:5 allocated:1 specially:1 comment:1 db:2 flow:1 effectiveness:1 integer:1 ee:2 feedforward:2 easy:1 enough:1 iterate:1 forthcoming:1 architecture:8 topology:3 perfectly:1 suboptimal:1 prototype:5 cn:1 motivated:1 penalty:1 speech:10 jilt:2 cocktail:7 listed:1 amount:2 mid:1 hardware:1 reduced:1 generate:1 notice:1 estimated:3 track:1 vol:6 nrmse:5 group:2 four:1 rissanen:7 achieving:1 drawn:1 ce:1 graph:1 fuse:1 year:1 forgiving:1 inverse:1 striking:1 place:1 family:4 reporting:1 reader:1 pursuing:1 separation:6 decision:1 comparable:1 layer:2 fl:1 followed:3 annual:1 bp:11 inclusive:1 x2:1 speed:1 vecchi:1 performing:1 expanded:1 gould:1 according:1 poor:1 smaller:2 describes:1 son:1 modification:2 restricted:1 taken:1 previously:1 remains:1 mechanism:1 committee:1 needed:1 know:1 studying:1 available:1 autoregression:1 rewritten:1 permit:1 hierarchical:2 appropriate:1 ochiai:1 gelatt:1 alternative:3 voice:3 original:3 move:1 pollution:1 added:2 question:2 strategy:2 parametric:1 gradient:1 distance:2 separate:2 simulated:5 sci:1 discriminant:1 reason:3 length:3 code:1 index:1 ratio:3 difficult:1 expense:1 negative:1 design:9 implementation:2 perform:1 neuron:3 modem:7 observation:3 purdue:4 predication:1 jak:1 descent:1 january:1 extended:1 discovered:1 arbitrary:2 sequent:1 david:1 pair:1 required:6 extensive:1 connection:6 functioned:1 wah:2 manoel:1 hour:1 nip:1 trans:4 address:1 able:1 recurring:1 below:2 perception:1 dynamical:1 lpc:1 oft:1 biasing:1 ivakhnenko:1 including:1 memory:1 explanation:1 difficulty:1 natural:1 representing:1 lao:1 created:1 prior:1 literature:1 contributing:1 fully:1 mixed:3 generation:1 limitation:2 digital:2 degree:1 consistent:2 fruit:1 principle:1 plotting:1 excitatory:1 elsewhere:1 loa:1 last:1 free:2 soon:1 guide:1 understand:1 taking:1 distributed:1 overcome:1 dimension:1 xn:1 axi:1 cumulative:2 made:3 adaptive:1 party:7 ec:1 far:1 hypersurface:1 reconstructed:2 ml:1 global:1 overfitting:1 automatica:1 search:11 channel:2 transfer:4 nature:2 ca:2 mol:1 complex:1 main:1 multilayered:1 noise:1 allowed:1 ny:1 xl:1 third:1 list:1 decay:1 ton:1 explored:1 bivariate:1 sequential:1 duffy:2 kx:1 led:1 explore:2 tenorio:15 subfamily:2 adjustment:1 contained:1 gender:1 radically:1 environmental:1 extracted:1 oct:1 identity:1 presentation:3 careful:1 change:2 hard:1 included:1 called:2 pas:1 accepted:1 select:2 people:1 evaluate:1 phenomenon:1 handling:2 |
1,817 | 2,650 | Newscast EM
Wojtek Kowalczyk
Department of Computer Science
Vrije Universiteit Amsterdam
The Netherlands
[email protected]
Nikos Vlassis
Informatics Institute
University of Amsterdam
The Netherlands
[email protected]
Abstract
We propose a gossip-based distributed algorithm for Gaussian mixture
learning, Newscast EM. The algorithm operates on network topologies
where each node observes a local quantity and can communicate with
other nodes in an arbitrary point-to-point fashion. The main difference
between Newscast EM and the standard EM algorithm is that the M-step
in our case is implemented in a decentralized manner: (random) pairs
of nodes repeatedly exchange their local parameter estimates and combine them by (weighted) averaging. We provide theoretical evidence and
demonstrate experimentally that, under this protocol, nodes converge exponentially fast to the correct estimates in each M-step of the EM algorithm.
1
Introduction
Advances in network technology, like peer-to-peer networks on the Internet or sensor networks, have highlighted the need for efficient ways to deal with large amounts of data that
are distributed over a set of nodes. Examples are financial data reported on the Internet,
weather data observed by a set of sensors, etc. In particular, in many data mining applications we are interested in learning a global model from such data, like a probability
distribution or a clustering of the data, without first transferring all the data to a central
repository. Ideally, we would like to have a fully decentralized algorithm that computes
and disseminates aggregates of the data, with minimal processing and communication requirements and good fault-tolerant behavior.
A recent development in distributed systems technology is the use of gossip-based models
of computation [1, 2, 3]. Roughly, in a gossip-based protocol each node repeatedly contacts some other node at random and the two nodes exchange information. Gossip-based
protocols are very simple to implement, while they enjoy strong performance guarantees
as a result of randomization. Their use in data mining and machine learning applications is
currently finding inroads [4, 5].
In this paper we propose a gossip-based, fully decentralized implementation of the
Expectation-Maximization (EM) algorithm for Gaussian mixture learning [6]. Our algorithm, which we call ?Newscast EM?, assumes a set of data {xi } that are drawn independently from a common Gaussian mixture and are distributed over the nodes of a network
(one data point per node). Newscast EM utilizes a gossip-based protocol in its M-step to
learn a global Gaussian mixture model p(x) from the data. The main idea is to perform the
M-step in a number of cycles. Each node starts with a local estimate of the model parameters. Then, in every cycle, each node contacts some other node that is chosen at random
from a list of known nodes, and the two nodes replace their local model estimates by their
(weighted) averages. As we show below, under such a protocol the (erroneous) local models of the individual nodes converge exponentially fast to the (correct) global model in each
M-step of the algorithm.
Our approach is fundamentally different from other distributed exact implementations of
the EM algorithm that resort on global broadcasting [7] or routing trees [8]. In the latter,
for instance, data sufficient statistics are propagated through a spanning tree in the network,
combined with an incremental learning scheme as in [9]. A disadvantage of that approach
is that only one node is carrying out computations at any time step, whereas in Newscast
EM all nodes are running the same protocol in parallel. This results in a batch M-step that
has average runtime at most logarithmic in the number of nodes, as we will see next.
2
Gaussian mixtures and the EM algorithm
A k-component Gaussian mixture for a random vector x ? IRd is defined as the convex
combination
k
X
p(x) =
?s p(x|s)
(1)
s=1
?d/2
of k Gaussian densities p(x|s) = (2?)
|Cs |?1/2 exp[?(x ? ms )> Cs?1 (x ? ms )/2],
each parameterized by its mean ms and covariance matrix Cs . The components of the
mixture are indexed by the random variable s that takes values from 1 to k, and ? s = p(s)
defines a discrete prior distribution over the components. Given a set {x1 , . . . , xn } of
independent and identically distributed samples from p(x), the learning task is to estimate
k
the parameter vector
Pn ? = {?s , ms , Cs }s=1 of the k components that maximizes the loglikelihood L = i=1 log p(xi ; ?). Throughout we assume that the likelihood function is
bounded from above (e.g., by placing appropriate bounds on the components covariance
matrices).
Maximization of the data log-likelihood L can be carried out by the EM algorithm [6]
which can be seen as iteratively maximizing a lower bound of L [9]. This bound F is
a function of the current mixture parameters ? and a set of ?responsibility? distributions
{qi (s)}, i = 1, . . . , n, where each qi (s) corresponds to a data point xi and defines an
arbitrary discrete distribution over s. This lower bound is given by:
F=
n X
k
X
i=1 s=1
qi (s) log p(xi , s; ?) ? log qi (s) .
(2)
In the E-step of the EM algorithm, the responsibility qi (s) for each point xi is set to the
Bayes posterior p(s|xi ) given the parameters found in the previous step. In the M-step we
solve for the unknown parameters of the mixture by maximizing F for fixed qi (s). This
yields the following updates:
Pn
Pn
Pn
qi (s)xi x>
i
i=1 qi (s)
i=1 qi (s)xi
?s =
,
ms =
,
Cs = i=1
?ms m>
s . (3)
n
n?s
n?s
Note that the main operation of the M-step is averaging: ?s is the average of qi (s), ms
is the average of products qi (s)xi (divided by ?s ), and the covariance matrix Cs is the
>
average of matrices qi (s)xi x>
i (divided by ?s and decreased by ms ms ). This observation
is essential for the proposed algorithm, as we will shortly see.
3
Newscast computing and averaging
The proposed distributed EM algorithm for Gaussian mixture learning relies on the use of
the Newscast protocol for distributed computing [3]. Newscast is a gossip-based protocol
that applies to networks where arbitrary point-to-point communication between nodes is
possible, and it involves repeated data exchange between nodes using randomization: with
constant frequency each node contacts some other node at random, and the two nodes
exchange application-specific data as well as caches with addresses of other nodes. The
protocol is very robust, scalable, and simple to implement?its Java implementation is
only a few kBytes of code and can run on small network-enabled computing devices such
as mobile phones, PDAs, or sensors.
As with other gossip-based protocols [2], Newscast can be used for computing the mean
of a set of values that are distributed over a network. Suppose that values v1 , . . . , vn are
stored in the nodes of a network, one value per node. Moreover
Pn suppose that each node
knows the addresses of all other nodes. To compute ? = n1 i=1 vi , each node i initially
sets ?i = vi as its local estimate of ?, and then runs the following protocol for a number of
cycles:
Uniform Newscast (for node i)
1. Contact a node j = f (i) that is chosen uniformly at random from 1, . . . , n.
2. Nodes i and j update their estimates as follows: ?0i = ?0j = (?i + ?j )/2.
For the purpose of analysis we will assume that in each cycle every node initiates a single
contact (but in practice the algorithm can be fully asynchronous). Note that the mean of
the local estimates {?i } is always the correct mean ?, while for their variance holds:
Lemma 1. In each cycle of uniform Newscast the variance of the local estimates drops on
1
the average by factor ?, with ? ? 2?
.
e
Pn
Proof. 1 Let ?t = i=1 (?i ? ?)2 be the unnormalized variance of the local estimates ?i
at cycle t. Suppose, without loss of generality, that within cycle t nodes initiate contacts in
the order 1, 2, . . . , n. The new variance after node?s 1 contact is:
? + ?
2
1
f (1)
?1 = ?t ? (?1 ? ?)2 ? (?f (1) ? ?)2 + 2
??
(4)
2
1
1
(5)
= ?t ? (?1 ? ?)2 ? (?f (1) ? ?)2 + (?1 ? ?)(?f (1) ? ?).
2
2
Taking expectation over f , and using the fact that P [f (i) = j] = n1 for all i and j, gives:
n
1 X
1
1
1
(?j ? ?)2 = 1 ?
? ? (?1 ? ?)2 . (6)
E[?1 |?t = ?] = ? ? (?1 ? ?)2 ?
2
2n j=1
2n
2
After n such exchanges, the variance ?t+1 is on the average:
n
1 n
1 X
1 n?i
E[?t+1 |?t = ?] = 1 ?
??
1?
(?i ? ?)2 .
2n
2 i=1
2n
Bounding the term (1 ?
1 n?i
2n )
by (1 ?
1 n
2n )
finally gives:
1 n
?
E[?t+1 |?t = ?] ?
1?
?? ? .
2
2n
2 e
1
(7)
1
See [3] for an alternative proof of the same bound.
(8)
Thus after t cycles of uniform Newscast,
? the original variance ?0 of the local estimates is
reduced on the average to ?t ? ?0 /(2 e)t . The fact that the variance drops at an exponential rate means that the nodes learn the correct average very fast. Indeed, using Chebyshev?s
inequality Pt [|?i ? ?| ? ?] ? ?t /(n?2 ) we see that for any ? > 0, the probability that
some node makes an estimation error larger than ? is dropping exponentially fast with the
number of cycles t. In particular, we can derive a bound on the number of cycles that are
needed in order to guarantee with high probability that all nodes know the correct answer
with some specific accuracy:
Theorem 1. With probability 1 ? ? , after d0.581(log n + 2 log ? + 2 log 1? + log 1? )e cycles
of uniform Newscast holds maxi |?i ? ?| ? ?, for any ? > 0 and data variance ? 2 .
?
Proof. Using Lemma 1 and the fact that ?0 = n? 2 , we obtain E[?t ] ? n? 2 /(2 e)t .
?
2
2
Setting ? = log( n?
?2 ? )/ log(2 e) we obtain E[?? ] ? ? ?. Using Markov inequality, with
2
probability at least 1 ? ? holds ?? ? ? . Therefore, since ?? is the sum of local terms, for
each of them must hold |?i ? ?| ? ?. It is straightforward to show by induction over ? that
the same inequality will hold for any time ? 0 > ? .
For example, for unit-variance data and a network with n = 104 nodes we need 49 cycles
to guarantee with probability 95% that each node is within 10?10 from the correct answer.
Note that in uniform Newscast, each node in the network is assumed to know the addresses
of all other nodes, and therefore can choose in each cycle one node uniformly at random
to exchange data with. In practice, however, each node can only have a limited cache of
addresses of other nodes. In this case, the averaging algorithm is modified as follows:
Non-uniform Newscast (for node i)
1. Contact a node j = f (i) that is appropriately chosen from i?s local cache.
2. Nodes i and j update their estimates as follows: ?0i = ?0j = (?i + ?j )/2.
3. Nodes i and j update their caches appropriately.
Step 3 implements a ?membership management? schedule which effectively defines a dynamically changing random graph topology over the network. In our experiments we
adopted the protocol of [10] which roughly operates as follows. Each entry k in node?s i
cache contains an ?age? attribute that indicates the number of cycles that have been elapsed
since node k created that entry. In step 1 above, node i contacts the node j with the largest
age from i?s cache, and increases by one the age of all other entries in i?s cache. Then
node i exchanges estimates with node j as in step 2. In step 3, both nodes i and j select a
random subset of their cache entries and mutually exchange them, filling empty slots and
discarding self-pointers and duplicates. Finally node i creates an entry with i?s address in
it and age zero, which is added in j?s cache. The resulting protocol is particularly effective
and, as we show in the experiments below, in some cases it even outperforms the uniform
Newscast. We refer to [10] for more details.
4
The Newscast EM algorithm
Newscast EM (NEM) is a gossip-based distributed implementation of the EM algorithm for
Gaussian mixture learning, that applies to the following setting. We are given a set of data
{xi } that are distributed over the nodes of a network (one data point per node). The data
are assumed independent samples from a common k-component Gaussian mixture p(x)
with (unknown) parameters ? = {?s , ms , Cs }ks=1 . The task is to learn the parameters of
the mixture with maximum likelihood in a decentralized manner: that is, all learning steps
should be performed locally at the nodes, and they should involve as little communication
as possible.
The NEM algorithm is a direct application of the averaging protocol of Section 3 for estimating the parameters ? of p(x) using the EM updates (3). The E-step of NEM is identical
to the E-step of the standard EM algorithm, and it can be performed by all nodes in parallel. The novel characteristic of NEM is the M-step which is implemented as a sequence
of gossip-based cycles: At the beginning of each M-step, each node i starts with a local
estimate ?i of the ?correct? parameter vector ? (correct according to EM and for the current
EM iteration). Then, in every cycle, each node contacts some other node at random, and
the two nodes replace their local estimates ?i by their (weighted) averages. At the end of
the M-step each node has converged (within machine precision) to the correct parameter ?.
To simplify notation, we will denote by ?i = {?si , msi , C?si } the local estimates of node i
for the parameters of component s at any point of the algorithm. The parameter C?si is
defined such that Csi = C?si ? msi m>
si . The complete algorithm, which runs identically
and in parallel for each node, is as follows:
Newscast EM (for node i)
1. Initialization. Set qi (s) to some random positive value and then normalize all qi (s) to
sum to 1 over all s.
2. M-step. Initialize i?s local estimates for each component s as follows: ?si = qi (s),
msi = xi , C?si = xi x>
i . Then repeat for ? cycles:
a. Contact a node j = f (i) from i?s local cache.
b. Nodes i and j update their local estimates for each component s as follows:
?si + ?sj
,
2
?si msi + ?sj msj
m0si = m0sj =
,
?si + ?sj
?si C?si + ?sj C?sj
0
0
.
C?si
= C?sj
=
?si + ?sj
0
0
?si
= ?sj
=
(9)
(10)
(11)
c. Nodes i and j update their caches appropriately.
3. E-step. Compute new responsibilities qi (s) = p(s|xi ) for each component s using the
M-step estimates ?si , msi , and Csi = C?si ? msi m>
si .
4. Loop. Go to step 2, unless a stopping criterion is satisfied that involves the parameter
estimates themselves or the energy F.2
A few observations about the algorithm are in order. First note that both the initialization
of the algorithm (step 1) as well as the E-step are completely local to each node. Similarly,
a stopping criterion involving the parameter estimates can be implemented locally if each
node caches its estimates from the previous EM-iteration. The M-step involves a total of
k[1 + d + d(d + 1)/2] averages, for each one of the k components and for dimensionality d, which are computed with the Newscast protocol. Given that all nodes agree on the
number ? of Newscast cycles in the M-step, and assuming that ? is large enough to guarantee convergence to the correct parameter estimates, the complete NEM algorithm can be
performed identically and in parallel by all nodes in the network.
It is easy to see that at any cycle of an M-step, and for any component s, the weighted
2
Note that F is a sum of local terms, and thus it can also be computed using the same protocol.
averages over all nodes of the local estimates are always the EM-correct estimates, i.e.,
Pn
i=1 ?si msi
P
= ms
(12)
n
i=1 ?si
and similarly for the C?si . Moreover, note that the weighted averages of the msi in (10)
and the C?si in (11), with weights given by (9), can be written as unweighted averages of
the corresponding products ?si msi and ?si C?si . In other words, each local estimate can be
written as the ratio of two local estimates that converge to the correct values at the same
exponential rate (as shown in the previous section). The above observations establish the
following:
Theorem 2. In every M-step of Newscast EM, each node converges exponentially fast to
the correct parameter estimates for each component of the mixture.
Similarly, the number of cycles ? for each M-step can be chosen according to Theorem 1.
However, note that in every M-step each node has to wait ? cycles before its local estimates
have converged, and only then can it use these estimates in a new next E-step. We describe
here a modification of NEM that allows a node to run a local E-step before its M-step has
converged. This ?partial? NEM algorithm is based on the following ?self-correction? idea:
instead of waiting until the M-step converges, after a small number of cycles each node runs
a local E-step, adjusts its responsibilities, and propagates appropriate corrections through
the network.
Such a scheme additionally requires that each node caches its responsibilities from the
previous E-step, denoted by q?i (s). The only modification is in the initialization of the Mstep: instead of fully resetting the local estimates as in step 2 above, a node makes the
following corrections to its current estimates ?si , msi , C?si for each component s:
0
?si
= ?si + qi (s) ? q?i (s),
m0si
0
C?si
(13)
0
q?i (s)]}/?si
,
= {msi ?si + xi [qi (s) ?
0
= {C?si ?si + xi x>
?i (s)]}/?si
.
i [qi (s) ? q
(14)
(15)
After these corrections, the Newscast averaging protocol is executed for a number of cycles (smaller than the number ? of the ?full? NEM). These corrections may increase the
variance of the local estimates, but in most cases the corresponding increase of variance is
relatively small. This results in speed-ups (often as large as 10-fold), however guaranteed
convergence is hard to establish.3
5
Experiments
To get an insight into the behavior of the presented algorithms we ran several experiments
using a Newscast simulator.4 In Fig. 1 we demonstrate the the performance of uniform
and non-uniform Newscast in typical averaging tasks involving zero-mean unit-variance
data. In Fig. 1(left) we plot the variance reduction rate ? (mean and one standard deviation
for 50 runs) as a function of the number of cycles, for averaging problems involving n =
105 data. Note
? that in uniform Newscast the observed rate is always below the derived
bound 1/(2 e) ? 0.303 from Lemma 1. Moreover note that in non-uniform Newscast the
variance drops faster than in uniform Newscast. This is due to the fact that the dynamic
cache exchange scheme of [10] results in in-degree network distributions that are very
peaked around the cache size. In practice this means that on the average each node is
3
This would require, for instance, that individual nodes have estimates of the total variance over
the network, which is not obvious how it can be done.
4
Available from http://www.cs.vu.nl/?steen/globesoul/sim.tgz
42.5
Uniform Newscast
Non?uniform (cache 20)
Uniform Newscast
Non?uniform (cache 20)
42
0.3
Cycles for convergence
Variance reduction rate
0.305
0.295
0.29
41
40.5
40
39.5
0.285
0
41.5
5
10
15
20
25
Number of cycles
30
35
40
39 3
10
4
5
10
10
6
10
Number of nodes
Figure 1: (Left) Variance reduction rate of uniform and non-uniform Newscast, in averaging tasks involving n = 105 nodes. (Right) Number of cycles to achieve convergence
within ? = 10?10 for unit-variance datasets of various sizes.
equally often contacted to by other nodes in each cycle of the protocol. We also observed
that the variance reduction rate is on the average unaffected by the network size, while
larger networks result in smaller deviations. For n = 8 ? 105 , for instance, the standard
deviation is half the one plotted above.
In Fig. 1(right) we plot the number of cycles that are required to achieve model accuracy at
all nodes within ? = 10?10 as a function of the network size. Note that all observed quantities are below the derived bound of Theorem 1, while non-uniform Newscast performs
slightly better than uniform Newscast.
We also ran experiments involving synthetic data drawn from Gaussian mixtures of different number of data points, where we observed results essentially identical to those obtained
by the standard (centralized) EM. We also performed some experiments with the ?partial?
NEM, where it turned out that in most cases we could obtain the same model accuracy with
a much smaller number of cycles (5?10 times than the ?full? NEM), but in some cases the
algorithm did not converge.
6
Summary and extensions
We presented Newscast EM, a distributed gossip-based implementation of the EM algorithm for learning Gaussian mixture models. Newscast EM applies on networks where
each one of a (large) number of nodes observes a local quantity, and can communicate with
other nodes in a point-to-point fashion. The algorithm utilizes a gossip-based protocol in
its M-step to learn a global Gaussian mixture model from the data: each node starts with
a local estimate of the parameters of the mixture and then, for a number of cycles till convergence, pairs of nodes repeatedly exchange their local parameter estimates and combine
them by (weighted) averaging. Newscast EM implements a batch M-step that has average runtime logarithmic in the network size. We believe that gossip-based protocols like
Newscast can be used in several other algorithms that learn models from distributed data.
Several extensions of the algorithm are possible. Here we have assumed that each node
in the network observes one data point. We can easily generalize this to situations where
each node observes (and stores) a collection of points, like in [8]. On the other hand, if the
locally observed data are too many, one may consider storing only some sufficient statistics of these data locally, and appropriately bound the energy F in each iteration to get a
convergent EM algorithm [11]. Another interesting extension is to replace the averaging
step 2 of uniform and non-uniform Newscast with weighted averaging (for some choice of
weights), and study the variance reduction rate in this case. Another interesting problem is
when the E-step cannot be performed locally at a node but it requires distributing some information over the network. This could be the case, for instance, when each node observes
only a few elements of a vector-valued quantity while, for instance, all nodes together observe the complete sample. We note that if the component models factorize, several useful
quantities can be computed by averaging in the log domain. Finally, it would be interesting
to investigate the applicability of the Newscast protocol in problems involving distributed
inference/learning in more general graphical models [12].
Acknowledgments
We want to thank Y. Sfakianakis for helping in the experiments, T. Pylak for making his
Newscast simulator publicly available, and D. Barber, Z. Ghahramani, and J.J. Verbeek for
their comments. N. Vlassis is supported by PROGRESS, the embedded systems research
program of the Dutch organization for Scientific Research NWO, the Dutch Ministry of
Economic Affairs and the Technology Foundation STW, project AES 5414.
References
[1] R. Karp, C. Schindelhauer, S. Shenker, and B. Vo? cking. Randomized rumour spreading. In
Proc. 41th IEEE Symp. on Foundations of Computer Science, Redondo Beach, CA, November
2000.
[2] D. Kempe, A. Dobra, and J. Gehrke. Gossip-based computation of aggregate information. In
Proc. 44th IEEE Symp. on Foundations of Computer Science, Cambridge, MA, October 2003.
[3] M. Jelasity, W. Kowalczyk, and M. van Steen. Newscast computing. Technical report, Dept. of
Computer Science, Vrije Universiteit Amsterdam, 2003. IR-CS-006.
[4] C. C. Moallemi and B. Van Roy. Distributed optimization in adaptive networks. In S. Thrun,
L. Saul, and B. Sch?olkopf, editors, Advances in Neural Information Processing Systems 16.
MIT Press, Cambridge, MA, 2004.
[5] D. Kempe and F. McSherry. A decentralized algorithm for spectral analysis. In Proc. 36th ACM
Symp. on Theory of Computing, Chicago, IL, June 2004.
[6] A. P. Dempster, N. M. Laird, and D. B. Rubin. Maximum likelihood from incomplete data via
the EM algorithm. J. Roy. Statist. Soc. B, 39:1?38, 1977.
[7] G. Forman and B. Zhang. Distributed data clustering can be efficient and exact. ACM SIGKDD
Explorations, 2(2):34?38, 2000.
[8] R. D. Nowak. Distributed EM algorithms for density estimation and clustering in sensor networks. IEEE Trans. on Signal Processing, 51(8):2245?2253, August 2003.
[9] R. M. Neal and G. E. Hinton. A view of the EM algorithm that justifies incremental, sparse, and
other variants. In M. I. Jordan, editor, Learning in graphical models, pages 355?368. Kluwer
Academic Publishers, 1998.
[10] S. Voulgaris, D. Gavidia, and M. van Steen. Inexpensive membership management for unstructured P2P overlays. Journal of Network and Systems Management, 2005. To appear.
[11] J. R. J. Nunnink, J. J. Verbeek, and N. Vlassis. Accelerated greedy mixture learning. In Proc.
Belgian-Dutch Conference on Machine Learning, Brussels, Belgium, January 2004.
[12] M. A. Paskin and C. E. Guestrin. Robust probabilistic inference in distributed systems. In Proc.
20th Int. Conf. on Uncertainty in Artificial Intelligence, Banff, Canada, July 2004.
| 2650 |@word repository:1 steen:3 covariance:3 reduction:5 contains:1 outperforms:1 current:3 si:35 must:1 written:2 chicago:1 mstep:1 drop:3 plot:2 update:7 half:1 greedy:1 device:1 intelligence:1 beginning:1 affair:1 kbytes:1 pointer:1 node:99 banff:1 zhang:1 direct:1 contacted:1 combine:2 symp:3 manner:2 indeed:1 roughly:2 themselves:1 behavior:2 simulator:2 little:1 cache:17 project:1 estimating:1 bounded:1 moreover:3 maximizes:1 notation:1 finding:1 guarantee:4 every:5 runtime:2 unit:3 enjoy:1 appear:1 positive:1 before:2 schindelhauer:1 local:31 initialization:3 k:1 dynamically:1 limited:1 acknowledgment:1 vu:2 practice:3 implement:4 java:1 weather:1 ups:1 word:1 wait:1 get:2 cannot:1 www:1 maximizing:2 straightforward:1 go:1 independently:1 convex:1 unstructured:1 adjusts:1 insight:1 financial:1 enabled:1 his:1 pt:1 suppose:3 exact:2 element:1 roy:2 particularly:1 observed:6 cycle:31 observes:5 csi:2 ran:2 dempster:1 pda:1 ideally:1 dynamic:1 carrying:1 creates:1 completely:1 easily:1 various:1 fast:5 effective:1 describe:1 artificial:1 aggregate:2 peer:2 larger:2 solve:1 valued:1 loglikelihood:1 statistic:2 highlighted:1 laird:1 sequence:1 propose:2 product:2 turned:1 loop:1 till:1 achieve:2 normalize:1 olkopf:1 nunnink:1 convergence:5 empty:1 requirement:1 incremental:2 converges:2 derive:1 progress:1 sim:1 strong:1 soc:1 implemented:3 c:10 involves:3 correct:13 attribute:1 exploration:1 routing:1 exchange:10 require:1 randomization:2 extension:3 helping:1 correction:5 hold:5 around:1 exp:1 belgium:1 purpose:1 estimation:2 proc:5 spreading:1 currently:1 nwo:1 largest:1 gehrke:1 weighted:7 mit:1 sensor:4 gaussian:13 always:3 modified:1 pn:7 mobile:1 karp:1 derived:2 june:1 likelihood:4 indicates:1 sigkdd:1 inference:2 stopping:2 membership:2 transferring:1 initially:1 interested:1 denoted:1 development:1 initialize:1 kempe:2 beach:1 identical:2 placing:1 filling:1 peaked:1 report:1 fundamentally:1 duplicate:1 few:3 simplify:1 individual:2 vrije:2 n1:2 organization:1 centralized:1 mining:2 investigate:1 mixture:19 wojtek:2 nl:3 mcsherry:1 nowak:1 partial:2 moallemi:1 belgian:1 unless:1 tree:2 indexed:1 incomplete:1 plotted:1 theoretical:1 minimal:1 instance:5 disadvantage:1 maximization:2 applicability:1 deviation:3 entry:5 subset:1 uniform:22 too:1 reported:1 stored:1 answer:2 synthetic:1 combined:1 density:2 randomized:1 probabilistic:1 informatics:1 together:1 central:1 satisfied:1 management:3 choose:1 conf:1 resort:1 int:1 vi:2 performed:5 view:1 responsibility:5 universiteit:2 bayes:1 start:3 parallel:4 p2p:1 il:1 publicly:1 accuracy:3 ir:1 variance:20 characteristic:1 resetting:1 yield:1 generalize:1 disseminates:1 unaffected:1 converged:3 inexpensive:1 energy:2 frequency:1 obvious:1 proof:3 propagated:1 dimensionality:1 schedule:1 dobra:1 done:1 generality:1 until:1 hand:1 defines:3 scientific:1 believe:1 iteratively:1 neal:1 deal:1 self:2 unnormalized:1 m:11 criterion:2 complete:3 demonstrate:2 vo:1 performs:1 novel:1 common:2 redondo:1 exponentially:4 shenker:1 kluwer:1 refer:1 cambridge:2 similarly:3 etc:1 posterior:1 recent:1 phone:1 store:1 inequality:3 fault:1 seen:1 ministry:1 guestrin:1 nikos:1 converge:4 signal:1 july:1 full:2 d0:1 technical:1 faster:1 academic:1 divided:2 equally:1 qi:19 verbeek:2 scalable:1 involving:6 variant:1 essentially:1 expectation:2 dutch:3 iteration:3 whereas:1 want:1 decreased:1 publisher:1 appropriately:4 sch:1 comment:1 jordan:1 call:1 identically:3 enough:1 easy:1 topology:2 economic:1 idea:2 chebyshev:1 tgz:1 distributing:1 ird:1 repeatedly:3 useful:1 involve:1 netherlands:2 amount:1 locally:5 statist:1 reduced:1 http:1 overlay:1 per:3 discrete:2 dropping:1 waiting:1 drawn:2 changing:1 v1:1 graph:1 sum:3 run:6 parameterized:1 uncertainty:1 communicate:2 throughout:1 vn:1 utilizes:2 bound:9 internet:2 guaranteed:1 convergent:1 fold:1 speed:1 relatively:1 department:1 according:2 inroad:1 brussels:1 combination:1 smaller:3 slightly:1 em:34 modification:2 making:1 mutually:1 agree:1 needed:1 know:3 initiate:2 end:1 adopted:1 available:2 operation:1 decentralized:5 observe:1 kowalczyk:2 appropriate:2 spectral:1 batch:2 alternative:1 shortly:1 original:1 assumes:1 clustering:3 running:1 graphical:2 ghahramani:1 establish:2 contact:11 added:1 quantity:5 thank:1 thrun:1 barber:1 spanning:1 induction:1 aes:1 assuming:1 code:1 msi:11 ratio:1 executed:1 october:1 stw:1 implementation:5 unknown:2 perform:1 observation:3 markov:1 datasets:1 november:1 january:1 situation:1 vlassis:4 communication:3 hinton:1 arbitrary:3 august:1 canada:1 pair:2 required:1 elapsed:1 forman:1 address:5 trans:1 below:4 program:1 scheme:3 technology:3 created:1 carried:1 prior:1 embedded:1 fully:4 loss:1 interesting:3 age:4 foundation:3 degree:1 sufficient:2 propagates:1 rubin:1 editor:2 storing:1 summary:1 repeat:1 supported:1 asynchronous:1 institute:1 saul:1 taking:1 sparse:1 distributed:18 van:3 xn:1 unweighted:1 computes:1 collection:1 adaptive:1 sj:8 global:5 tolerant:1 assumed:3 xi:16 factorize:1 additionally:1 learn:5 robust:2 ca:1 protocol:21 domain:1 uva:1 did:1 main:3 bounding:1 repeated:1 x1:1 fig:3 gossip:14 fashion:2 precision:1 exponential:2 theorem:4 erroneous:1 msj:1 specific:2 discarding:1 paskin:1 maxi:1 list:1 evidence:1 essential:1 effectively:1 justifies:1 broadcasting:1 logarithmic:2 amsterdam:3 applies:3 corresponds:1 relies:1 acm:2 ma:2 slot:1 nem:10 replace:3 experimentally:1 hard:1 typical:1 operates:2 uniformly:2 averaging:13 lemma:3 total:2 newscast:42 select:1 latter:1 accelerated:1 dept:1 |
1,818 | 2,651 | Multiple Relational Embedding
Roland Memisevic
Department of Computer Science
University of Toronto
[email protected]
Geoffrey Hinton
Department of Computer Science
University of Toronto
[email protected]
Abstract
We describe a way of using multiple different types of similarity relationship to learn a low-dimensional embedding of a dataset. Our method
chooses different, possibly overlapping representations of similarity by
individually reweighting the dimensions of a common underlying latent
space. When applied to a single similarity relation that is based on Euclidean distances between the input data points, the method reduces to
simple dimensionality reduction. If additional information is available
about the dataset or about subsets of it, we can use this information to
clean up or otherwise improve the embedding. We demonstrate the potential usefulness of this form of semi-supervised dimensionality reduction on some simple examples.
1
Introduction
Finding a representation for data in a low-dimensional Euclidean space is useful both for
visualization and as prelude to other kinds of data analysis. The common goal underlying the many different methods that accomplish this task (such as ISOMAP [1], LLE [2],
stochastic neighbor embedding [3] and others) is to extract the usually small number of
factors that are responsible for the variability in the data. In making the underlying factors
explicit, these methods help to focus on the kind of variability that is important and provide
representations that make it easier to interpret and manipulate the data in reasonable ways.
Most dimensionality reduction methods are unsupervised, so there is no way of guiding
the method towards modes of variability that are of particular interest to the user. There
is also no way of providing hints when the true underlying factors are too subtle to be
discovered by optimizing generic criteria such as maximization of modeled variance in
PCA, or preservation of local geometry in LLE. Both these difficulties can be alleviated by
allowing the user to provide more information than just the raw data points or a single set
of pairwise similarities between data points.
As an example consider images of faces. Nonlinear methods have been shown to find
embeddings that nicely reflect the variability in the data caused by variation in face identity,
pose, position, or lighting effects. However, it is not possible to tell these methods to
extract a particular single factor for the purpose of, say intelligent image manipulation or
pose identification, because the extracted factors are intermingled and may be represented
simultaneously across all latent space dimensions.
Here, we consider the problem of learning a latent representation for data based on knowl-
edge that is provided by a user in the form of several different similarity relations. Our
method, multiple relational embedding (MRE), finds an embedding that uses a single latent data representation, but weights the available latent space dimensions differently to
allow the latent space to model the multiple different similarity relations. By labeling a
subset of the data according to the kind of variability one is interested in, one can encourage the model to reserve a subset of the latent dimensions for this kind of variability.
The model, in turn, returns a ?handle? to that latent space in the form of a corresponding
learned latent space metric. Like stochastic neighbor embedding, MRE can also be derived
as a simplification of Linear Relational Embedding[4].
1.1
Related work
The problem of supplementing methods for unsupervised learning with ?side-information?
in order to influence their solutions is not new and many different approaches have been
suggested. [5], for example, describes a way to inform a PCA model by encouraging it to
preserve a user-defined grouping structure; [6] consider the problem of extracting exactly
two different kinds of factors, which they denote ?style? and ?content?, by using bilinear
models; more recently, [7] and [8] took a quite different approach to informing a model.
They suggest pre-processing the input data by learning a metric in input space that makes
the data respect user defined grouping constraints.
Our approach differs from these and other methods in two basic ways. First, in all the
methods mentioned above, the side-information has to be defined in terms of equivalence
constraints. That is, a user needs to define a grouping structure for the input data by informing the model which data-points belong together. Here, we consider a rather different
approach, where the side-information can be encoded in the form of similarity relations.
This allows arbitrary continuous degrees of freedom to constrain the low-dimensional embeddings. Second, our model can deal with several, possibly conflicting, kinds of sideinformation. MRE dynamically ?allocates? latent space dimensions to model different
user-provided similarity relations. So inconsistent relations are modeled in disjoint subspaces, and consistent relations can share dimensions. This scheme of sharing the dimensions of a common latent space is reminiscent of the INDSCAL method [9] that has been
popular in the psychometric literature.
A quite different way to extend unsupervised models has recently been introduced by [10]
and [11], where the authors propose ways to extract common factors that underlie two
or more different datasets, with possibly different dimensionalities. While these methods rely on a supervision signal containing information about correspondences between
data-points in different datasets, MRE can be used to discover correspondences between
different datasets using almost no pre-defined grouping constraints.
2
Multiple Relational Embedding
In the following we derive MRE as an extension to stochastic neighbor embedding (SNE).
Let X denote the matrix of latent space elements arranged column-wise, and let ? 2 be some
real-valued neighborhood variance or ?kernel bandwidth?. SNE finds a low-dimensional
representation for a set of input data points y i (i = 1, . . . , N ) by first constructing a similarity matrix P with entries
exp(? ?12 ky i ? y j k2 )
Pij := P
1
i
k 2
k exp(? ? 2 ky ? y k )
(1)
and then minimizing (w.r.t. the set of latent space elements xi (i = 1, . . . , N )) the mismatch
between P and the corresponding latent similarity matrix Q(X) defined by
exp(?kxi ? xj k2 )
.
Qij (X) := P
i
k 2
k exp(?kx ? x k )
(2)
The (row-) normalization of both matrices arises from SNE?s probabilistic formulation in
which the (i, j)th entry of P and Q is interpreted as the probability that the ith data-point
will pick the j th point as its neighbor (in observable and latent space, respectively). The
mismatch is defined as the sum of Kullback-Leibler-divergences between the respective
rows [3].
Our goal is to extend SNE so that it learns latent data representations that not only approximate the input space distances well, but also reflect additional characteristics of the input
data that one may be interested in. In order to accommodate these additional characteristics, instead of defining a single similarity-matrix that is based on Euclidean distances in
data space, we define several matrices P c , (c = 1, . . . , C), each of which encodes some
known type of similarity of the data. Proximity in the Euclidean data-space is typically one
of the types of similarity that we use, though it can easily be omitted. The additional types
of similarity may reflect any information that the user has access to about any subsets of
the data provided the information can be expressed as a similarity matrix that is normalized
over the relevant subset of the data.
At first sight, a single latent data representation seems to be unsuitable to accommodate the
different, and possibly incompatible, properties encoded in a set of P c -matrices. Since our
goal, however, is to capture possibly overlapping relations, we do use a single latent space
and in addition we define a linear transformation Rc of the latent space for each of the C
different similarity-types that we provide as input. Note that this is equivalent to measuring
distances in latent space using a different Mahalanobis metric for each c corresponding to
the matrix Rc T Rc .
In order to learn the transformations Rc from the data along with the set of latent representations X we consider the loss function
X
E(X) =
E c (X),
(3)
c
where we define
1 X c
E c (X) :=
P log
N i,j ij
Pijc
Qcij
!
and
Qcij := Qij (Rc X).
(4)
Note that in the case of C = 1, R1 = I (and fixed) and P 1 defined as in Eq. (1) this
function simplifies to the standard SNE objective function. One might consider weighting
the contribution of each similarity-type using some weighting factor ?c . We found that
the solutions are rather robust with regard to different sets of ?c and weighted all error
contributions equally in our experiments.
As indicated above, here we consider diagonal R-matrices only, which simply amounts to
using a rescaling factor for each latent space dimension. By allowing each type of similarity
to put a different scaling factor on each dimension the model allows similarity relations
that ?overlap? to share dimensions. Completely unrelated or ?orthogonal? relations can be
encoded by using disjoint sets of non-zero scaling factors.
The gradient of E(X) w.r.t. a single latent space element xl takes a similar form to the
gradient of the standard SNE objective function and is given by
?E(X)
2 XX c
=
(Pil + Plic ? Qcli ? Qcil ) Rc T Rc (xl ? xi ),
(5)
l
?x
N c i
2
0
?2
?4
?6 REucl
0.5
0.5
0
0
?0.5
1
?0.5
0.5
1
0.5
0
0
0.5
0
?0.5
?0.5
?1
2
3
0
?2
?4
0
?1
3
2
0.5
1
?0.5
?0.5
2
?1
?6 RClass
?1
1
Figure 1: Embedding of images of rotated objects. Left: SNE, right: MRE. Latent representatives are colored on a gray-scale corresponding to angle of rotation in the original
images. The rightmost plots show entries on the diagonals of latent space transformations
REucl and RClass .
the gradient w.r.t. to a single entry of the diagonal of Rc reads
?
?
?E(X)
2 c ?X X c
j
R
Pij ? Qcij (xil ? xl )2 ? ,
=
c
?Rll
N ll i j
(6)
where xil denotes the lth component of the ith latent representative.
As an illustrative example we ran MRE on a set of images from the Columbia object images
library (COIL) [12]. The dataset contains (128 ? 128)-dimensional gray-scale images of
different objects that vary only by rotation, i.e. by a single degree of freedom. We took
three subsets of images depicting toy-cars, where each subset corresponds to one of three
different kinds of toy-cars, and embedded the first 30 images of each of these subsets in a
three-dimensional space. We used two similarity relations: The first, P Eucl , corresponds
to the standard SNE objective; the second, P Class , is defined as a block diagonal matrix
1
that contains homogeneous blocks of size 30 ? 30 with entries ( 30
) and models class
membership, i.e. we informed the model using the information that images depicting the
same object class belong together.
We also ran standard SNE on the same dataset1 . The results are depicted in figure 1. While
SNE?s unsupervised objective to preserve Euclidean distances leads to a representation
where class-membership is intermingled with variability caused by object rotation (leftmost plot), in the MRE approximation the contribution of class-membership is factored out
and represented in a separate dimension (next plot). This is also reflected in the entries
on the diagonal of the corresponding R-matrices, depicted in the two right-most plots.
RClass is responsible for representing class membership and can do so using just a single
dimension. REucl on the other hand makes use of all dimensions to some degree, reflecting
the fact that the overall variability in ?pixel-space? depends on class-membership, as well
as on other factors (here mainly rotation). Note that with the variability according to class1
For training we set ? 2 manually to 5 ? 107 for both SNE and MRE and initialized all entries in X
and the diagonals of all Rc with small normally distributed values. In all experiments we minimized
the loss function defined in Eq. (3) using Carl Rasmussens? matlab function ?minimize? for 200
iterations (simple gradient descent worked equally well, but was much slower).
membership factored out, the remaining two dimensions capture the rotational degree of
freedom very cleanly.
2.1
Partial information
In many real world situations there might be side-information available only for a subset
of the data-points, because labelling a complete dataset could be too expensive or for other
reasons impossible. A partially labelled dataset can in that case still be used to provide
a hint about the kind of variability that one is interested in. In general, since the corresponding transformation Rc provides a way to access the latent space that represents the
desired similarity-type, a partially labelled dataset can be used to perform a form of supervised feature extraction in which the labelled data is used to specify a kind of feature ?by
example?. It is straightforward to modify the model to deal with partially labelled data.
For each type of similarity c that is known to hold for a subset containing N c examples,
the corresponding P c -matrix references only this subset of the complete dataset and is thus
an N c ? N c -matrix. To keep the latent space elements not corresponding to this subset
unaffected by this error contribution, we can define for each c an index set I c containing
just the examples referenced by P c and rewrite the loss for that type of similarity as
!
c
X
P
1
ij
E c (X) := c
.
(7)
Pijc log
N
Qcij
c
i,j?I
3
Experiments
3.1
Learning correspondences between image sets
In extending the experiment described in section 2 we trained MRE to discover correspondences between sets of images, in this case with different dimensionalities. We picked 20
successive images from one object of the COIL dataset described above and 28 images
(112 ? 92 pixels) depicting a person under different viewing angles taken from the UMIST
dataset[13]. We chose this data in order to obtain two sets of images that vary in a ?similar?
or related way. Note that, because the datasets have different dimensionalities, here it is not
possible to define a single relation describing Euclidean distance between all data-points.
Instead we constructed two relations P Coil and P Umist (for both we used Eq. (1) with ? 2
set as in the previous experiment), with corresponding index-sets I Coil and I Umist containing the indices of the points in each of the two datasets. In addition we constructed one
class-membership relation in the same way as before and two identical relations P 1 and
P 2 that take the form of a 2 ? 2-matrix filled with entries 12 . Each of the corresponding
index sets I 1 and I 2 points to two images (one from each dataset) that represent the end
points of the rotational degree of freedom, i.e. to the first and the last points if we sort the
data according to rotation (see figure 2, left plot). These similarity types are used to make
sure that the model properly aligns the representations of the two different datasets. Note
that the end points constitute the only supervision signal; we did not use any additional
information about the alignment of the two datasets.
After training a two-dimensional embedding2 , we randomly picked latent representatives
of the COIL images and computed reconstructions of corresponding face images using a
kernel smoother (i.e. as a linear combination of the face images with coefficients based on
latent space distances). In order to factor out variability corresponding to class membership
we first multiplied all latent representatives by the inverse of R class . (Note that such a
strategy will in general blow up the latent space dimensions that do not represent class
membership, as the corresponding entries in Rclass may contain very small values. The
2
Training was done using 500 iterations with a setup as in the previous experiment.
Figure 2: Face reconstructions by alignment. Left: Side-information in form of two image
pairs in correspondence. Right: Reconstructions of face images from randomly chosen cat
images.
kernel smoother consequently requires a very large kernel bandwidth, with the net effect
that the latent representation effectively collapses in the dimensions that correspond to class
membership ? which is exactly what we want.) The reconstructions, depicted in the right
plot of figure 2, show that the model has captured the common mode of variability.
3.2
Supervised feature extraction
To investigate the ability of MRE to perform a form of ?supervised feature extraction? we
used a dataset of synthetic face images that originally appeared in [1]. The face images
vary according to pose (two degrees of freedom) and according to the position of a lighting
source (one degree of freedom). The corresponding low-dimensional parameters are available for each data-point. We computed an embedding with the goal of obtaining features
that explicitly correspond to these different kinds of variability in the data.
We labelled a subset of 100 out of the total of 698 data-points with the three mentioned
degrees of freedom in the following way: After standardizing the pose and lighting parameters so that they were centered and had unit variance, we constructed three corresponding
similarity matrices (P Pose1 , P Pose2 , P Lighting ) for a randomly chosen subset of 100 points
using Eq. (1) and the three low-dimensional parameter sets as input data. In addition we
used a fourth similarity relation P Ink , corresponding to overall brightness or ?amount of
ink?, by constructing for each image a corresponding feature equal to the sum of its pixel
intensities and then defining the similarity matrix as above. We set the bandwidth parameter ? 2 to 1.0 for all of these similarity-types3 . In addition we constructed the standard SNE
relation P Eucl (defined for all data-points) using Eq. (1) with ? 2 set4 to 100.
We initialized the model as before and trained for 1000 iterations of ?minimize? to find
an embedding in a four-dimensional space. Figure 3 (right plot) shows the learned latent
space metrics corresponding to the five similarity-types. Obviously, MRE devotes one
dimension to each of the four similarity-types, reflecting the fact that each of them describes
a single one-dimensional degree of freedom that is barely correlated with the others. Dataspace similarities in contrast are represented using all dimensions. The plots on the left
of figure 3 show the embedding of the 598 unlabelled data-points. The top plot shows the
embedding in the two dimensions in which the two ?pose?-metrics take on their maximal
values, the bottom plot shows the dimensions in which the ?lighting?- and ?ink?-metric take
on their maximal values. The plots show that MRE generalizes over unlabeled data: In each
dimension the unlabeled data is clearly arranged according to the corresponding similarity
type, and is arranged rather randomly with respect to other similarity types. There are
a few correlations, in particular between the first pose- and the ?ink?-parameter, that are
inherent in the dataset, i.e. the data does not vary entirely independently with respect to
these parameters. These correlations are also reflected in the slightly overlapping latent
3
This is certainly not an optimal choice, but we found the solutions to be rather robust against
changes in the bandwidth, and this value worked fine.
4
See previous footnote.
0.3
5
REucl
0.2
0
x1
0.1
?5
1
0
5
?0.1
2
3
4
3
4
3
4
2
3
4
2
3
4
RLights
0
?0.2
?5
1
?0.3
?0.2
?0.1
0
0.1
0.2
0.3
x4
5
2
RPose2
0
0.3
?5
1
0.2
5
x3
0.1
2
RPose1
0
0
?5
1
?0.1
Ink
5
?0.2
R
0
?0.3
?5
?0.2
?0.1
0
0.1
x2
0.2
0.3
1
Figure 3: Left: Embedding of faces images that were not informed about their lowdimensional parameters. For a randomly chosen subset of these (marked with a circle),
the original images are shown next to their latent representatives. Right: Entries on the
diagonals of five latent space transformations.
space weight sets. MRE gets the pose-embedding wrong for a few very dark images that
are apparently too far away in the data space to be associated with the correct labeled datapoints.
4
Conclusions
We introduced a way to embed data in a low-dimensional space using a set of similarity
relations. Our experiments indicate that the informed feature extraction that this method facilitates will be most useful in cases where conventional dimensionality reduction methods
fail because of their completely unsupervised nature. Although we derived our approach
as an extension to SNE, it should be straightforward to apply the same idea to other dimensionality reduction methods.
Acknowledgements: Roland Memisevic is supported by a Government of Canada Award.
Geoffrey Hinton is a fellow of CIAR and holds a CRC chair. This research was also supported by grants from NSERC and CFI.
References
[1] Joshua B. Tenenbaum, Vin de Silva, and John C. Langford. A global geometric framework for
nonlinear dimensionality reduction. Science, pages 2319?2323, 2000.
[2] S.T. Roweis and L. K. Saul. Nonlinear dimensionality reduction by locally linear embedding.
Science, 290, 2000.
[3] Geoffrey Hinton and Sam Roweis. Stochastic neighbor embedding. In Advances in Neural
Information Processing Systems 15, pages 833?840. MIT Press, 2003.
[4] A. Paccanaro and G. E. Hinton. Learning hierarchical structures with linear relational embedding. In Advances in Neural Information Processing Systems 14, Cambridge, MA, 2002. MIT
Press.
[5] David Cohn. Informed projections. In Advances in Neural Information Processing Systems 15,
pages 849?856. MIT Press, 2003.
[6] Joshua B. Tenenbaum and William T. Freeman. Separating style and content with bilinear
models. Neural Computation, 12(6):1247?1283, 2000.
[7] Eric P. Xing, Andrew Y. Ng, Michael I. Jordan, and Stuart Russell. Distance metric learning with application to clustering with side-information. In Advances in Neural Information
Processing Systems 15, pages 505?512. MIT Press, Cambridge, MA, 2003.
[8] Michinari Momma Tijl De Bie and Nello Cristianini. Efficiently learning the metric using sideinformation. In Proc. of the 14th International Conference on Algorithmic Learning Theory,
2003.
[9] J. Douglas Carroll and Jih-Jie Chang. Analysis of individual differences in multidimensional
scaling via an n-way generalization of ?eckart-young? decomposition. Psychometrika, 35(3),
1970.
[10] J. H. Ham, D. D. Lee, and L. K. Saul. Learning high dimensional correspondences from low
dimensional manifolds. In In Proceedings of the ICML 2003 Workshop on The Continuum from
Labeled to Unlabeled Data in Machine Learning and Data Mining, pages 34?41, Washington,
D.C., 2003.
[11] Jakob J. Verbeek, Sam T. Roweis, and Nikos Vlassis. Non-linear cca and pca by alignment of local models. In Advances in Neural Information Processing Systems 16. MIT Press, Cambridge,
MA, 2004.
[12] S. A. Nene, S. K. Nayar, and H. Murase. Columbia object image library (coil-20). Technical
report, 1996.
[13] Daniel B Graham and Nigel M Allinson. Characterizing virtual eigensignatures for general
purpose face recognition. 163, 1998.
| 2651 |@word momma:1 seems:1 cleanly:1 decomposition:1 pick:1 brightness:1 accommodate:2 reduction:7 contains:2 daniel:1 rightmost:1 bie:1 reminiscent:1 john:1 class1:1 plot:11 ith:2 colored:1 provides:1 toronto:4 successive:1 five:2 rc:10 along:1 constructed:4 qij:2 pairwise:1 freeman:1 encouraging:1 supplementing:1 psychometrika:1 provided:3 discover:2 underlying:4 unrelated:1 xx:1 what:1 kind:10 interpreted:1 informed:4 finding:1 transformation:5 fellow:1 multidimensional:1 exactly:2 k2:2 wrong:1 normally:1 underlie:1 unit:1 grant:1 before:2 local:2 modify:1 referenced:1 tijl:1 bilinear:2 might:2 chose:1 equivalence:1 dynamically:1 collapse:1 responsible:2 block:2 differs:1 x3:1 cfi:1 alleviated:1 projection:1 pre:2 suggest:1 get:1 unlabeled:3 put:1 influence:1 impossible:1 equivalent:1 conventional:1 rll:1 straightforward:2 independently:1 sideinformation:2 factored:2 datapoints:1 embedding:20 handle:1 variation:1 user:8 homogeneous:1 us:1 carl:1 element:4 expensive:1 recognition:1 labeled:2 bottom:1 capture:2 eckart:1 russell:1 ran:2 mentioned:2 ham:1 cristianini:1 trained:2 rewrite:1 eric:1 completely:2 easily:1 differently:1 represented:3 cat:1 describe:1 tell:1 labeling:1 intermingled:2 neighborhood:1 quite:2 encoded:3 valued:1 say:1 otherwise:1 ability:1 obviously:1 net:1 took:2 propose:1 reconstruction:4 lowdimensional:1 maximal:2 relevant:1 roweis:3 ky:2 r1:1 extending:1 xil:2 rotated:1 object:7 help:1 derive:1 andrew:1 pose:7 ij:2 eq:5 c:2 murase:1 indicate:1 correct:1 stochastic:4 centered:1 viewing:1 virtual:1 crc:1 government:1 generalization:1 extension:2 hold:2 proximity:1 exp:4 algorithmic:1 reserve:1 vary:4 continuum:1 omitted:1 purpose:2 proc:1 knowl:1 individually:1 weighted:1 mit:5 qcij:4 clearly:1 sight:1 rather:4 jih:1 derived:2 focus:1 properly:1 mainly:1 contrast:1 membership:10 typically:1 relation:18 interested:3 pixel:3 overall:2 michinari:1 equal:1 nicely:1 extraction:4 ng:1 manually:1 identical:1 represents:1 x4:1 stuart:1 unsupervised:5 icml:1 washington:1 minimized:1 report:1 others:2 intelligent:1 hint:2 few:2 inherent:1 randomly:5 simultaneously:1 preserve:2 divergence:1 individual:1 set4:1 geometry:1 william:1 freedom:8 interest:1 investigate:1 mining:1 certainly:1 alignment:3 eucl:2 edge:1 encourage:1 partial:1 allocates:1 respective:1 orthogonal:1 filled:1 euclidean:6 initialized:2 desired:1 circle:1 column:1 umist:3 measuring:1 maximization:1 subset:15 entry:10 usefulness:1 too:3 nigel:1 accomplish:1 kxi:1 chooses:1 synthetic:1 person:1 international:1 memisevic:2 probabilistic:1 lee:1 michael:1 together:2 reflect:3 containing:4 possibly:5 style:2 return:1 rescaling:1 toy:2 potential:1 de:2 blow:1 standardizing:1 coefficient:1 devotes:1 caused:2 explicitly:1 depends:1 picked:2 apparently:1 xing:1 sort:1 pil:1 vin:1 contribution:4 minimize:2 variance:3 characteristic:2 efficiently:1 correspond:2 eigensignatures:1 raw:1 identification:1 lighting:5 unaffected:1 footnote:1 nene:1 inform:1 sharing:1 aligns:1 against:1 associated:1 dataset:12 popular:1 car:2 dimensionality:10 subtle:1 reflecting:2 mre:14 originally:1 supervised:4 reflected:2 specify:1 arranged:3 formulation:1 though:1 done:1 prelude:1 just:3 correlation:2 langford:1 hand:1 cohn:1 nonlinear:3 overlapping:3 reweighting:1 mode:2 gray:2 indicated:1 effect:2 normalized:1 true:1 isomap:1 contain:1 read:1 leibler:1 deal:2 mahalanobis:1 ll:1 allinson:1 illustrative:1 criterion:1 leftmost:1 paccanaro:1 complete:2 demonstrate:1 silva:1 image:29 wise:1 recently:2 common:5 rotation:5 belong:2 extend:2 interpret:1 cambridge:3 had:1 access:2 similarity:34 supervision:2 carroll:1 optimizing:1 manipulation:1 joshua:2 captured:1 additional:5 nikos:1 signal:2 semi:1 preservation:1 multiple:5 smoother:2 reduces:1 technical:1 unlabelled:1 roland:3 manipulate:1 equally:2 award:1 verbeek:1 basic:1 metric:8 iteration:3 kernel:4 normalization:1 represent:2 addition:4 want:1 fine:1 source:1 sure:1 facilitates:1 inconsistent:1 jordan:1 extracting:1 embeddings:2 xj:1 bandwidth:4 simplifies:1 idea:1 ciar:1 pca:3 constitute:1 matlab:1 jie:1 useful:2 amount:2 dark:1 tenenbaum:2 locally:1 disjoint:2 four:2 douglas:1 clean:1 sum:2 angle:2 inverse:1 fourth:1 almost:1 reasonable:1 incompatible:1 scaling:3 graham:1 entirely:1 cca:1 simplification:1 correspondence:6 constraint:3 worked:2 constrain:1 x2:1 encodes:1 chair:1 department:2 according:6 combination:1 across:1 describes:2 slightly:1 sam:2 making:1 taken:1 visualization:1 turn:1 describing:1 fail:1 dataspace:1 end:2 available:4 generalizes:1 multiplied:1 apply:1 hierarchical:1 away:1 generic:1 slower:1 original:2 denotes:1 remaining:1 top:1 clustering:1 unsuitable:1 ink:5 objective:4 strategy:1 diagonal:7 gradient:4 subspace:1 distance:8 separate:1 separating:1 manifold:1 nello:1 reason:1 barely:1 modeled:2 relationship:1 index:4 providing:1 minimizing:1 rotational:2 setup:1 sne:13 perform:2 allowing:2 datasets:7 descent:1 defining:2 relational:5 hinton:5 variability:13 situation:1 vlassis:1 discovered:1 jakob:1 arbitrary:1 intensity:1 canada:1 introduced:2 david:1 pair:1 learned:2 conflicting:1 suggested:1 usually:1 mismatch:2 appeared:1 overlap:1 difficulty:1 rely:1 representing:1 scheme:1 improve:1 library:2 extract:3 columbia:2 literature:1 acknowledgement:1 geometric:1 embedded:1 loss:3 geoffrey:3 degree:9 pij:2 consistent:1 share:2 row:2 supported:2 last:1 rasmussen:1 side:6 lle:2 allow:1 neighbor:5 saul:2 face:10 characterizing:1 distributed:1 regard:1 dimension:21 world:1 dataset1:1 author:1 far:1 approximate:1 observable:1 kullback:1 keep:1 global:1 xi:2 continuous:1 latent:37 learn:2 nature:1 robust:2 correlated:1 obtaining:1 depicting:3 constructing:2 did:1 x1:1 psychometric:1 representative:5 position:2 guiding:1 explicit:1 xl:3 weighting:2 learns:1 young:1 embed:1 grouping:4 workshop:1 effectively:1 labelling:1 kx:1 easier:1 depicted:3 simply:1 expressed:1 nserc:1 partially:3 chang:1 corresponds:2 extracted:1 ma:3 coil:6 lth:1 goal:4 identity:1 marked:1 consequently:1 towards:1 informing:2 labelled:5 content:2 change:1 total:1 arises:1 nayar:1 |
1,819 | 2,652 | Conditional Random Fields for Object
Recognition
Ariadna Quattoni Michael Collins Trevor Darrell
MIT Computer Science and Artificial Intelligence Laboratory
Cambridge, MA 02139
{ariadna, mcollins, trevor}@csail.mit.edu
Abstract
We present a discriminative part-based approach for the recognition of
object classes from unsegmented cluttered scenes. Objects are modeled
as flexible constellations of parts conditioned on local observations found
by an interest operator. For each object class the probability of a given
assignment of parts to local features is modeled by a Conditional Random Field (CRF). We propose an extension of the CRF framework that
incorporates hidden variables and combines class conditional CRFs into
a unified framework for part-based object recognition. The parameters of
the CRF are estimated in a maximum likelihood framework and recognition proceeds by finding the most likely class under our model. The main
advantage of the proposed CRF framework is that it allows us to relax the
assumption of conditional independence of the observed data (i.e. local
features) often used in generative approaches, an assumption that might
be too restrictive for a considerable number of object classes.
1 Introduction
The problem that we address in this paper is that of learning object categories from supervised data. Given a training set of n pairs (xi , yi ), where xi is the ith image and yi is the
category of the object present in xi , we would like to learn a model that maps images to
object categories. In particular, we are interested in learning to recognize rigid objects such
as cars, motorbikes, and faces from one or more fixed view-points.
The part-based models we consider represent images as sets of patches, or local features,
which are detected by an interest operator such as that described in [4]. Thus an image
xi can be considered to be a vector {xi,1 , . . . , xi,m } of m patches. Each patch xi,j has
a feature-vector representation ?(xi,j ) ? Rd ; the feature vector might capture various
features of the appearance of a patch, as well as features of its relative location and scale.
This scenario presents an interesting challenge to conventional classification approaches in
machine learning, as the input space xi is naturally represented as a set of feature-vectors
{?(xi,1 ), . . . , ?(xi,m )} rather than as a single feature vector. Moreover, the local patches
underlying the local feature vectors may have complex interdependencies: for example,
they may correspond to different parts of an object, whose spatial arrangement is important
to the classification task.
The most widely used approach for part-based object recognition is the generative model
proposed in [1]. This classification system models the appearance, spatial relations and
co-occurrence of local parts. One limitation of this framework is that to make the model
computationally tractable one has to assume the independence of the observed data (i.e.,
local features) given their assignment to parts in the model. This assumption might be too
restrictive for a considerable number of object classes made of structured patterns.
A second limitation of generative approaches is that they require a model P (xi,j |hi,j ) of
patches xi,j given underlying variables hi,j (e.g., hi,j may be a hidden variable in the
model, or may simply be yi ). Accurately specifying such a generative model may be challenging ? in particular in cases where patches overlap one another, or where we wish to
allow a hidden variable hi,j to depend on several surrounding patches. A more direct approach may be to use a feature vector representation of patches, and to use a discriminative
learning approach. We follow an approach of this type in this paper.
Similar observations concerning the limitations of generative models have been made in
the context of natural language processing, in particular in sequence-labeling tasks such as
part-of-speech tagging [7, 5, 3] and in previous work on conditional random fields (CRFs)
for vision [2]. In sequence-labeling problems for NLP each observation xi,j is typically
the j?th word for some input sentence, and hi,j is a hidden state, for example representing
the part-of-speech of that word. Hidden Markov models (HMMs), a generative approach,
require a model of P (xi,j |hi,j ), and this can be a challenging task when features such as
word prefixes or suffixes are included in the model, or where hi,j is required to depend
directly on words other than xi,j . This has led to research on discriminative models for sequence labeling such as MEMM?s [7, 5] and conditional random fields (CRFs)[3]. A strong
argument for these models as opposed to HMMs concerns their flexibility in terms of representation, in that they can incorporate essentially arbitrary feature-vector representations
?(xi,j ) of the observed data points.
We propose a new model for object recognition based on Conditional Random Fields. We
model the conditional distribution p(y|x) directly. A key difference of our approach from
previous work on CRFs is that we make use of hidden variables in the model. In previous
work on CRFs (e.g., [2, 3]) each ?label? yi is a sequence hi = {hi,1 , hi,2 , . . . , hi,m } of
labels hi,j for each observation xi,j . The label sequences are typically taken to be fully
observed on training examples. In our case the labels yi are unstructured labels from some
fixed set of object categories, and the relationship between yi and each observation xi,j is
not clearly defined. Instead, we model intermediate part-labels hi,j as hidden variables in
the model. The
P model defines conditional probabilities P (y, h | x), and hence indirectly
P (y | x) = h P (y, h | x), using a CRF. Dependencies between the hidden variables
h are modeled by an undirected graph over these variables. The result is a model where
inference and parameter estimation can be carried out using standard graphical model algorithms such as belief propagation [6].
2 The Model
2.1 Conditional Random Fields with Hidden Variables
Our task is to learn a mapping from images x to labels y. Each y is a member of a set Y
of possible image labels, for example, Y = {background, car}. We take each image x
to be a vector of m ?patches? x = {x1 , x2 , . . . , xm }.1 Each patch xj is represented by a
feature vector ?(xj ) ? Rd . For example, in our experiments each xj corresponds to a patch
that is detected by the feature detector in [4]; section [3] gives details of the feature-vector
representation ?(xj ) for each patch. Our training set consists of labeled images (xi , yi ) for
i = 1 . . . n, where each yi ? Y, and each xi = {xi,1 , xi,2 , . . . , xi,m }. For any image x
we also assume a vector of ?parts? variables h = {h1 , h2 , . . . , hm }. These variables are
not observed on training examples, and will therefore form a set of hidden variables in the
1
Note that the number of patches m can vary across images, and did vary in our experiments. For
convenience we use notation where m is fixed across different images; in reality it will vary across
images but this leads to minor changes to the model.
model. Each hj is a member of H where H is a finite set of possible parts in the model.
Intuitively, each hj corresponds to a labeling of xj with some member of H. Given these
definitions of image-labels y, images x, and part-labels h, we will define a conditional
probabilistic model:
e?(y,h,x;?)
.
(1)
P (y, h | x, ?) = P
?(y 0 ,h,x;?)
y 0 ,h e
Here ? are the parameters of the model, and ?(y, h, x; ?) ? R is a potential function
parameterized by ?. We will discuss the choice of ? shortly. It follows that
P ?(y,h,x;?)
X
e
(2)
P (y | x, ?) =
P (y, h | x, ?) = P h ?(y0 ,h,x;?) .
y 0 ,h e
h
Given a new test image x, and parameter values ?? induced from a training example, we
will take the label for the image to be arg maxy?Y P (y | x, ?? ). Following previous work
on CRFs [2, 3], we use the following objective function in training the parameters:
X
1
(3)
L(?) =
log P (yi | xi , ?) ? 2 ||?||2
2?
i
The first term in Eq. 3 is the log-likelihood of the? data. The
? second term is the log of a
Gaussian prior with variance ? 2 , i.e., P (?) ? exp 2?1 2 ||?||2 . We will use gradient ascent
to search for the optimal parameter values, ?? = arg max? L(?), under this criterion.
We now turn to the definition of the potential function ?(y, h, x; ?). We assume an undirected graph structure, with the hidden variables {h1 , . . . , hm } corresponding to vertices
in the graph. We use E to denote the set of edges in the graph, and we will write (j, k) ? E
to signify that there is an edge in the graph between variables hj and hk . In this paper we
assume that E is a tree.2 We define ? to take the following form:
?(y, h, x; ?) =
m X
X
j=1
fl1 (j, y, hj , x)?l1 +
l
X X
(j,k)?E
fl2 (j, k, y, hj , hk , x)?l2
(4)
l
where fl1 , fl2 are functions defining the features in the model, and ?l1 , ?l2 are the components
of ?. The f 1 features depend on single hidden variable values in the model, the f 2 features
can depend on pairs of values. Note that ? is linear in the parameters ?, and the model in
Eq. 1 is a log-linear model. Moreover the features respect the structure of the graph, in that
no feature depends on more than two hidden variables hj , hk , and if a feature does depend
on variables hj and hk there must be an edge (j, k) in the graph E.
Assuming that the edges in E form a tree, and that ? takes the form in Eq. 4, then exact
methods exist for inference and parameter estimation in the model. This follows because
belief propagation [6] can be used to calculate the following quantities in O(|E||Y|) time:
X
?y ? Y, Z(y | x, ?) =
exp{?(y, h, x; ?)}
h
?y ? Y, j ? 1 . . . m, a ? H,
X
P (hj = a | y, x, ?) =
P (h | y, x, ?)
h:hj =a
?y ? Y, (j, k) ? E, a, b ? H,
P (hj = a, hk = b | y, x, ?) =
X
P (h | y, x, ?)
h:hj =a,hk =b
2
This will allow exact methods for inference and parameter estimation in the model, for example
using belief propagation. If E contains cycles then approximate methods, such as loopy beliefpropagation, may be necessary for inference and parameter estimation.
The first term Z(y | x, ?) is a partition function defined by a summation over
the h variables.
Terms of this form can be used to calculate P (y | x, ?) =
P
Z(y | x, ?)/ y0 Z(y 0 | x, ?). Hence inference?calculation of arg max P (y | x, ?)?
can be performed efficiently in the model. The second and third terms are marginal distributions over individual variables hj or pairs of variables hj , hk corresponding to edges in
the graph. The next section shows that the gradient of L(?) can be defined in terms of these
marginals, and hence can be calculated efficiently.
2.2 Parameter Estimation Using Belief Propagation
This section considers estimation of the parameters ?? = arg max L(?) from a training
sample, where L(?) is defined in Eq. 3. In our work we used a conjugate-gradient method
to optimize L(?) (note that due to the use of hidden variables, L(?) has multiple local
minima, and our method is therefore not guaranteed to reach the globally optimal point).
In this section we describe how the gradient of L(?) can be calculated efficiently. Consider
the likelihood term that is contributed by the i?th training example, defined as:
!
? P
?(yi ,h,xi ;?)
e
(5)
Li (?) = log P (yi | xi , ?) = log P h ?(y0 ,h,x ;?) .
i
y 0 ,h e
We first consider derivatives with respect to the parameters ?l1 corresponding to features
fl1 (j, y, hj , x) that depend on single hidden variables. Taking derivatives gives
?Li (?) X
??(yi , h, xi ; ?) X
??(y 0 , h, xi ; ?)
=
P (h | yi , xi , ?)
?
P (y 0 , h | xi , ?)
1
1
??l
??l
??l1
0
h
=
X
y ,h
P (h | yi , xi , ?)
=
fl1 (j, yi , hj , xi ) ?
P (hj = a |
X
P (y 0 , h | xi , ?)
y 0 ,h
j=1
h
X
m
X
yi , xi , ?)fl1 (j, yi , a, xi )
?
X
m
X
fl1 (j, y 0 , hj , xi )
j=1
0
P (hj = a, y | xi , ?)fl1 (j, y 0 , a, xi )
y 0 ,j,a
j,a
can be expressed in terms of components P (hj = a | xi , ?) and
It follows that ?L??i (?)
1
l
P (y | xi , ?), which can be calculated using belief propagation, provided that the graph E
forms a tree structure. A similar calculation gives
?Li (?)
??l2
X
=
P (hj = a, hk = b | yi , xi , ?)fl2 (j, k, yi , a, b, xi )
(j,k)?E,a,b
X
?
P (hj = a, hk = b, y 0 | xi , ?)fl2 (j, k, y 0 , a, b, xi )
y 0 ,(j,k)?E,a,b
hence ?Li (?)/??l2 can also be expressed in terms of expressions that can be calculated
using belief propagation.
2.3 The Specific Form of our Model
We now turn to the specific form for the model in this paper. We define
X
X
X
?(y, h, x; ?) =
?(xj ) ? ?(hj ) +
?(y, hj ) +
?(y, hj , hk )
j
j
(6)
(j,k)?E
Here ?(k) ? Rd for k ? H is a parameter vector corresponding to the k?th part label. The
inner-product ?(xj ) ? ?(hj ) can be interpreted as a measure of the compatibility between
patch xj and part-label hj . Each parameter ?(y, k) ? R for k ? H, y ? Y can be
interpreted as a measure of the compatibility between part k and label y. Finally, each
parameter ?(y, k, l) ? R for y ? Y, and k, l ? H measures the compatibility between an
edge with labels k and l and the label y. It is straightforward to verify that the definition in
Eq. 6 can be written in the same form as Eq. 4. Hence belief propagation can be used for
inference and parameter estimation in the model.
The patches xi,j in each image are obtained using the SIFT detector [4]. Each patch xi,j
is then represented by a feature vector ?(xi,j ) that incorporates a combination of SIFT and
relative location and scale features.
The tree E is formed by running a minimum spanning tree algorithm over the parts hi,j ,
where the cost of an edge in the graph between hi,j and hi,k is taken to be the distance
between xi,j and xi,k in the image. Note that the structure of E will vary across different
images. Our choice of E encodes our assumption that parts conditioned on features that are
spatially close are more likely to be dependent. In the future we plan to experiment with
the minimum spanning tree approach under other definitions of edge-cost. We also plan to
investigate more complex graph structures that involve cycles, which may require approximate methods such as loopy belief propagation for parameter estimation and inference.
3 Experiments
We carried out three sets of experiments on a number of different data sets.3 The first
two experiments consisted of training a two class model (object vs. background) to distinguish between a category from a single viewpoint and background. The third experiment
consisted of training a multi-class model to distinguish between n classes.
The only parameter that was adjusted in the experiments was the scale of the images upon
which the interest point detector was run. In particular, we adjusted the scale on the car
side data set: in this data set the images were too small and without this adjustment the
detector would fail to find a significant amount of features.
For the experiments we randomly split each data set into three separate data sets: training,
validation and testing. We use the validation data set to set the variance parameters ? 2 of
the gaussian prior.
3.1 Results
In figure 2.a we show how the number of parts in the model affects performance. In the case
of the car side data set, the ten-part model shows a significant improvement compared to
the five parts model while for the car rear data set the performance improvement obtained
by increasing the number of parts is not as significant. Figure 2.b shows a performance
comparison with previous approaches [1] tested on the same data set (though on a different
partition). We observe an improvement between 2 % and 5 % for all data sets.
Figures 3 and 4 show results for the multi-class experiments. Notice that random performance for the animal data set would be 25 % across the diagonal. The model exhibits
best performance for the Leopard data set, for which the presence of part 1 alone is a clear
predictor of the class. This shows again that our model can learn discriminative part distributions for each class. Figure 3 shows results for a multi-view experiment where the task
is two distinguish between two different views of a car and background.
3
The images were obtained from http://www.vision.caltech.edu/html-files/archive.html and the
car side images from http://l2r.cs.uiuc.edu/ cogcomp/Data/Car/. Notice, that since our algorithm
does not currently allow for the recognition of multiple instances of an object we test it on a partition
of the the training set in http://l2r.cs.uiuc.edu/ cogcomp/Data/Car/ and not on the testing set in that
site. The animals data set is a subset of Caltech?s 101 categories data set.
Figure 1: Examples of the most likely assignment of parts to features for the two class
experiments (car data set).
Data set
(a) Car Side
Car Rear
5 parts
94 %
91 %
10 parts
99 %
91.7 %
Data set
Car Side
Car Rear
(b) Face
Plane
Motorbike
Our Model
99 %
94.6 %
99 %
96 %
95 %
Others [1]
90.3 %
96.4 %
90.2 %
92.5 %
Figure 2: (a) Equal Error Rates for the car side and car rear experiments with different
number of parts. (b) Comparative Equal Error Rates.
Figure 1 displays the Viterbi labeling4 for a set of example images showing the most likely
assignment of local features to parts in the model. Figure 6 shows the mean and variance
of each part?s location for car side images and background images. The mean and variance
of each part?s location for the car side images were calculated in the following manner:
First we find for every image classified as class a the most likely part assignment under our
model. Second, we calculate the mean and variance of positions of all local features that
were assigned to the same part. Similarly Figure 5 shows part counts among the Viterbi
paths assigned to examples of a given class.
As can be seen in Figure 6 , while the mean location of a given part in the background
images and the mean location of the same part in the car images are very similar, the parts
in the car have a much tighter distribution which seems to suggest that the model is learning
the shape of the object.
As shown in Figure 5 the model has also learnt discriminative part distributions for each
class, for example the presence of part 1 seems to be a clear predictor for the car class. In
general part assignments seem to rely on a combination of appearance and relative location.
Part 1, for example, is assigned to wheel like patterns located on the left of the object.
4
This is the labeling h? = arg maxh P (h | y, x, ?) where x is an image and y is the label for
the image under the model.
Data set
Car Side
Car Rear
Precision
87.5 %
87.4 %
Recall
98 %
86.5 %
Figure 3: Precision and recall results for 3 class experiment.
Data set
Leopards
Llamas
Rhinos
Pigeons
Leopards
91 %
0%
0%
0%
Llamas
2%
50 %
40 %
30 %
Rhinos
0%
27 %
46 %
20 %
Pigeons
7%
23 %
14 %
50 %
Figure 4: Confusion table for 4 class experiment.
However, the parts might not carry semantic meaning. It appears that the model has learnt
a vocabulary of very general parts with significant variability in appearance and learns to
discriminate between classes by capturing the most likely arrangement of these parts for
each class.
In some cases the model relies more heavily on relative location than appearance because
the appearance information might not be very useful for discriminating between the two
classes. One of the reasons for this is that the detector produces a large number of false detections, making the appearance data too noisy for discrimination. The fact that the model
is able to cope with this lack of discriminating appearance information illustrates its flexible
data-driven nature. This can be a desirable model property of a general object recognition
system, because for some object classes appearance is the important discriminant (i.e., in
textured classes) while for others shape may be important (i.e., in geometrically constrained
classes).
One noticeable difference between our model and similar part-based models is that our
model learns large parts composed of small local features. This is not surprising given how
the part dependencies were built (i.e., through their position in minimum spanning tree):
the potential functions defined on pairs of hidden variables tend to smooth the allocation of
parts to patches.
1
3
3
8
4
5
8
76
5
4
Figure 5: Graph showing part counts for the background (left) and car side images (right)
4 Conclusions and Further Work
In this work we have presented a novel approach that extends the CRF framework by incorporating hidden variables and combining class conditional CRFs into an unified framework for object recognition. Similarly to CRFs and other maximum entropy models our
approach allows us to combine arbitrary observation features for training discriminative
classifiers with hidden variables. Furthermore, by making some assumptions about the
joint distribution of hidden variables one can derive efficient training algorithms based on
dynamic programming.
Background Shape
Car Shape
200
200
150
150
100
100
3
50
50
4
5
0
1
8
?50
?50
?100
?100
?150
?150
?200
?200
?150
?100
?50
0
(a)
3
5
0
50
100
150
200
?200
?200
?150
?100
?50
9
7
6
4
8
0
50
100
150
200
(b)
Figure 6: (a) Graph showing mean and variance of locations for the different parts for the
car side images; (b) Mean and variance of part locations for the background images.
The main limitation of our model is that it is dependent on the feature detector picking up
discriminative features of the object. Furthermore, our model might learn to discriminate
between classes based on the statistics of the feature detector and not the true underlying
data, to which it has no access. This is not a desirable property since it assumes the feature
detector to be consistent. As future work we would like to incorporate the feature detection
process into the model.
References
[1] R. Fergus, P. Perona,and A. Zisserman. Object class recognition by unsupervised scale-invariant
learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,volume 2, pages 264-271, June 2003.
[2] S. Kumar and M. Hebert. Discriminative random fields: A framework for contextual interaction
in classification. In IEEE Int Conference on Computer Vision,volume 2, pages 1150-1157, June
2003.
[3] J. Lafferty,A. McCallum and F. Pereira. Conditional random fields: Probabilistic models for
segmenting and labeling sequence data. In Proc. Int Conf. on Machine Learning, 2001.
[4] D. Lowe. Object Recognition from local scale-invariant features. In IEEE Int Conference on
Computer Vision, 1999.
[5] A. McCallum, D. Freitag, and F. Pereira. Maximum entropy markov models for information
extraction and segmentation. In ICML-2000, 2000.
[6] J. Pearl. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. Morgan Kaufmann,1988.
[7] A. Ratnaparkhi. A maximum entropy part-of-speech tagger. In EMNLP, 1996.
| 2652 |@word seems:2 carry:1 contains:1 prefix:1 contextual:1 surprising:1 must:1 written:1 partition:3 shape:4 v:1 alone:1 intelligence:1 generative:6 discrimination:1 plane:1 mccallum:2 ith:1 location:10 five:1 tagger:1 direct:1 consists:1 freitag:1 combine:2 manner:1 tagging:1 uiuc:2 multi:3 globally:1 increasing:1 provided:1 notation:1 moreover:2 underlying:3 interpreted:2 unified:2 finding:1 every:1 classifier:1 segmenting:1 local:13 path:1 might:6 specifying:1 challenging:2 co:1 hmms:2 testing:2 word:4 suggest:1 convenience:1 close:1 wheel:1 operator:2 context:1 optimize:1 conventional:1 map:1 www:1 crfs:8 straightforward:1 cluttered:1 unstructured:1 mcollins:1 heavily:1 exact:2 programming:1 recognition:12 located:1 labeled:1 observed:5 capture:1 calculate:3 cycle:2 dynamic:1 depend:6 upon:1 textured:1 joint:1 various:1 represented:3 surrounding:1 describe:1 artificial:1 detected:2 labeling:6 whose:1 widely:1 plausible:1 ratnaparkhi:1 relax:1 statistic:1 noisy:1 advantage:1 sequence:6 propose:2 interaction:1 product:1 combining:1 flexibility:1 darrell:1 produce:1 comparative:1 object:25 derive:1 minor:1 noticeable:1 eq:6 strong:1 c:2 require:3 tighter:1 summation:1 adjusted:2 extension:1 leopard:3 considered:1 exp:2 mapping:1 viterbi:2 vary:4 estimation:8 proc:1 label:17 currently:1 mit:2 clearly:1 gaussian:2 rather:1 hj:26 june:2 improvement:3 likelihood:3 hk:10 inference:8 dependent:2 rigid:1 suffix:1 rear:5 typically:2 hidden:19 relation:1 rhino:2 perona:1 interested:1 compatibility:3 arg:5 classification:4 flexible:2 html:2 among:1 plan:2 spatial:2 animal:2 constrained:1 marginal:1 field:8 equal:2 extraction:1 unsupervised:1 icml:1 future:2 others:2 intelligent:1 randomly:1 composed:1 recognize:1 individual:1 detection:2 interest:3 investigate:1 cogcomp:2 edge:8 necessary:1 tree:7 instance:1 assignment:6 loopy:2 cost:2 vertex:1 subset:1 predictor:2 too:4 dependency:2 learnt:2 discriminating:2 csail:1 probabilistic:3 picking:1 michael:1 again:1 opposed:1 emnlp:1 conf:1 derivative:2 li:4 potential:3 int:3 llama:2 depends:1 performed:1 view:3 h1:2 lowe:1 formed:1 variance:7 kaufmann:1 efficiently:3 correspond:1 accurately:1 classified:1 quattoni:1 detector:8 reach:1 trevor:2 definition:4 naturally:1 recall:2 car:26 segmentation:1 appears:1 supervised:1 follow:1 zisserman:1 though:1 furthermore:2 unsegmented:1 propagation:8 lack:1 defines:1 verify:1 consisted:2 true:1 hence:5 assigned:3 spatially:1 laboratory:1 semantic:1 criterion:1 crf:6 confusion:1 l1:4 reasoning:1 image:35 meaning:1 novel:1 volume:2 marginals:1 significant:4 cambridge:1 rd:3 similarly:2 language:1 access:1 maxh:1 driven:1 scenario:1 yi:19 caltech:2 seen:1 morgan:1 minimum:4 multiple:2 desirable:2 interdependency:1 fl1:7 smooth:1 calculation:2 concerning:1 vision:5 essentially:1 represent:1 background:9 signify:1 archive:1 ascent:1 file:1 induced:1 tend:1 undirected:2 member:3 incorporates:2 lafferty:1 seem:1 presence:2 intermediate:1 split:1 independence:2 xj:8 affect:1 inner:1 expression:1 speech:3 useful:1 clear:2 involve:1 amount:1 ten:1 category:6 http:3 exist:1 notice:2 estimated:1 write:1 key:1 graph:13 geometrically:1 run:1 parameterized:1 extends:1 patch:18 capturing:1 hi:16 guaranteed:1 distinguish:3 display:1 scene:1 x2:1 encodes:1 argument:1 kumar:1 structured:1 fl2:4 combination:2 conjugate:1 across:5 y0:3 making:2 maxy:1 intuitively:1 invariant:2 taken:2 computationally:1 discus:1 turn:2 fail:1 count:2 tractable:1 observe:1 indirectly:1 occurrence:1 shortly:1 motorbike:2 assumes:1 running:1 nlp:1 graphical:1 restrictive:2 objective:1 beliefpropagation:1 arrangement:2 quantity:1 diagonal:1 exhibit:1 gradient:4 distance:1 separate:1 considers:1 discriminant:1 spanning:3 reason:1 assuming:1 modeled:3 relationship:1 contributed:1 observation:6 markov:2 finite:1 defining:1 variability:1 arbitrary:2 pair:4 required:1 sentence:1 pearl:1 address:1 able:1 proceeds:1 pattern:3 xm:1 challenge:1 l2r:2 built:1 max:3 belief:8 overlap:1 natural:1 rely:1 representing:1 carried:2 hm:2 prior:2 l2:4 relative:4 fully:1 interesting:1 limitation:4 allocation:1 validation:2 h2:1 consistent:1 viewpoint:1 ariadna:2 hebert:1 side:11 allow:3 face:2 taking:1 calculated:5 vocabulary:1 made:2 cope:1 approximate:2 discriminative:8 xi:50 fergus:1 search:1 reality:1 table:1 learn:4 nature:1 complex:2 did:1 main:2 x1:1 site:1 precision:2 position:2 pereira:2 wish:1 third:2 learns:2 specific:2 sift:2 showing:3 constellation:1 concern:1 incorporating:1 false:1 conditioned:2 illustrates:1 entropy:3 led:1 pigeon:2 simply:1 likely:6 appearance:9 expressed:2 adjustment:1 corresponds:2 relies:1 ma:1 conditional:13 considerable:2 change:1 included:1 discriminate:2 collins:1 incorporate:2 tested:1 |
1,820 | 2,653 | A Topographic Support Vector Machine:
Classification Using Local Label Configurations
Johannes Mohr
Clinic for Psychiatry and Psychotherapy
Charit?e Medical School
and
Bernstein Center for Computational Neuroscience Berlin
10117 Berlin, Germany
Klaus Obermayer
Department of Electrical Engineering and Computer Science
Berlin University of Technology
and
Bernstein Center for Computational Neuroscience Berlin
10587 Berlin, Germany
[email protected], [email protected]
Abstract
The standard approach to the classification of objects is to consider the
examples as independent and identically distributed (iid). In many real
world settings, however, this assumption is not valid, because a topographical relationship exists between the objects. In this contribution we
consider the special case of image segmentation, where the objects are
pixels and where the underlying topography is a 2D regular rectangular
grid. We introduce a classification method which not only uses measured
vectorial feature information but also the label configuration within a topographic neighborhood. Due to the resulting dependence between the
labels of neighboring pixels, a collective classification of a set of pixels
becomes necessary. We propose a new method called ?Topographic Support Vector Machine? (TSVM), which is based on a topographic kernel
and a self-consistent solution to the label assignment shown to be equivalent to a recurrent neural network. The performance of the algorithm is
compared to a conventional SVM on a cell image segmentation task.
1
Introduction
The segmentation of natural images into semantically meaningful subdivisions can be considered as one or more binary pixel classification problems, where two classes of pixels are
characterized by some measurement data (features). For each binary problem the task is
to assign a set of new pixels to one of the two classes using a classifier trained on a set of
labeled pixels (training data).
In conventional classification approaches usually the assumption of iid examples is made,
so the classification result is determined solely by the measurement data. Natural images,
however, possess a topographic structure, in which there are dependencies between the
labels of topographic neighbors, making the data non-iid. Therefore, not only the measurement data, but also the labels of the topographic neighbors can be used in the classification
of a pixel. It has been shown for a number of problems that dependencies between instances can improve model accuracy. A Conditional Random Field approach approach has
been used for labeling text sequences by [1]. Combining this idea with local discriminative
models, in [2] a discriminative random field was used to model the dependencies between
the labels of image blocks in a probabilistic framework. A collective classification relational dependency network was used in [3] for movie box-office receipts prediction and
paper topic classification. The maximization of the per label margin of pairwise Markov
networks was applied in [4] to handwritten character recognition and collective hypertext
classification. There, the number of variables and constraints of the quadratic programming
problem was polynomial in the number of labels.
In this work, we propose a method which is also based on margin maximization and allows
the collective assignments of a large number of binary labels which have a regular grid
topography. In contrast to [4] the number of constraints and variables does not depend on
the number of labels. The method called topographic support vector machine (TSVM) is
based on the assumption that knowledge about the local label configuration can improve the
classification of a single data point. Consider as example the segmentation of a collection
of images depicting physical objects of similar shape, but high variability in gray level and
texture. In this case, the measurements are dissimilar, while the local label configurations
show high similarity.
Here, we apply the TSVM to the supervised bottom-up segmentation of microscopic images of Papanicolaou stained cervical cell nuclei from the CSSIP pap smear dataset 1 . Segmentation of these images is important for the detection of cervical cancer or precancerous
cells. The final goal is to use so-called malignancy associated changes (MACs), e.g. a
slight shift of the distribution of nuclear size not yet visual to the human observer, in order
to detect cancer at an early stage [5]. A previously used bottom-up segmentation approach
for this data using morphological watersheds was reported to have difficulties with weak
gradients and the presence of other large gradients adjacent to the target [5]. Top-down
methods like active contour models have successfully been used [6], but require heuristic
initialization and error correction procedures.
2
Classification using a Topographic Support Vector Machine
Let O = {o1 , ..., on } be a set of n sites on a 2D pixel-grid and G = {Go , o ? O} be
a neighborhood system for O, where Go is the set of neighbors of o and neighborhood
is defined by o 6? Go and o ? Gp ? p ? Go . For each pixel site oi from the set O,
a binary label yi ? {?1, +1} giving the class assignment is assumed to be known. To
simplify the notation, in the following we are going to make use of multi-indices written in
the form of vectors, referring to pairs of indices on a two-dimensional grid. We define the
neighborhood of order c as G c = {Gi , i ? O}; Gi = {k ? O : 0 < (k?i)2 ? c}. This way,
G 1 describes the first order neighborhood system (4 neighbors), G 2 the second order system
(8 neighbors), and so on. Each pixel site is characterized by some measurement vector. This
could for example be the vector of gray value intensities at a pixel site, the gray value patch
around a central pixel location, or the responses to a bank of linear or nonlinear filters (e.g.
Gabor coefficients). Using a training set composed of (possibly several) sets of pixel sites,
each accompanied by a set of measurement vectors X = {xi , ?i ? [1..n]} and a set of
1
Centre for Sensor Signal and Information Processing, University of Queensland
labels Y = {yi , ?i ? [1..n]} (e.g. a manually labeled image), the task of classification is
to assign class labels to a set of ? pixels sites U = {u1 , ..., u? } of an unlabeled image, for
? = {?
which a set of measurements X
xi , ?i ? [1..?]} is available. For the classification we
will use a support vector machine.
2.1
Support Vector Classification
In Support Vector Classification (SVC) methods ([7]), a kernel is used to solve a complex
classification task in a usually high-dimensional feature space via a separating hyperplane.
Results from statistical learning theory ([8]) show that maximizing the margin (the distance
of the closest data point to the hyperplane) leads to improved generalization abilities. In
practice, the optimal margin hyperplane can be obtained solving a quadratic programming
problem. Several schemes have been introduced to deal with noisy measurements via the
introduction of slack variables. In the following we will shortly review one such scheme,
the C-SVM, which is also later used in the experiments. For a canonical separating hyperplane (w, b) in a higher dimensional feature space H, to which the n variables x i are
mapped by ?(x), and n slack variables ?i the primal objective function of a C-SVM can
be formulated as
n
1
X
w
2 + C
min n
?i ,
(1)
w?H,??R
2
n i=1
subject to yi (wT ?(xi ) + b) ? 1 ? ?i ,
?i ? 0,
C > 0,
i = 1, ..., n.
In order to classify a new object h with unknown label, the following decision rule is
evaluated:
m
X
f (xh ) = sgn
?i yi K(xh , xi ) + b ,
(2)
i=1
where the sum runs over all m support vectors.
2.2
Topographic Kernel
We now assume that the label of each pixel is determined by both the measurement and
the set of labels of its topographic neighbors. We define a vector yGh where the labels of
the q topographic neighbors of the pixel h are concatenated in an arbitrary, but fixed order.
We propose a support vector classifier using an extended kernel, which in addition to the
measurement vector xh , also includes the vector yGh :
K(xh , xj , yGh , yGj ) = K1 (xh , xj ) + ? ? K2 (yGh , yGj ),
(3)
where ? is a hyper-parameter. Kernel K1 can be an arbitrary kernel working on the measurements. For kernel K2 an arbitrary dot-product kernel might be used. In the following
we restrict ourselves to a linear kernel (corresponding to the normalized Hamming distance
between the local label configurations)
K2 (yGh , yGj ) =
1
hyGh |yGj i,
q
(4)
where h...|...i denotes a scalar product. The kernel K2 defined in eq. (4) thus consists of a
dot-product between these vectors divided by their length. For a neighborhood G hc of order
c we obtain
X
1
K2 (yGh , yGj ) =
yh+s ? yj+s
(5)
q
?
|s|< c,s6=0
The linear kernel K2 in (4) takes on its maximum value, if the label configurations are
identical, and its lowest value if the label configuration is inverted.
2.3
Learning phase
If a SVM is trained using the topographic kernel (3), the topographic label configuration is
included in the learning process. The resulting support vectors will still contain the relevant
information about the measurements, but additionally the label neighborhood information
relevant for a good distinction of the classes.
2.4
Classification phase
In order to collectively classify a set of ? new pixel sites with unknown topographic label
configuration, we propose the following iterative approach to achieve a self-consistent solution to the classification problem. We denote the labels at step ? as yh (? ), ?h. At each
step ? new labels are assigned according to
m
X
yh (? ) = sgn
j=1
?j ? yv(j) ? K(xh , xv(j) , yGh (? ? 1), yGv(j) ) + b , ?h.
(6)
The sum runs over all m support vectors, whose indices on the 2D grid are denoted by
the vector v(j). Since initially the labels are unknown, we use at step ? = 0 the results
from a conventional support-vector machine (? = 0) as initialization for the labels. For the
following steps some estimates of the neighboring labels are available from the previous
iteration. Using this new topographic label information in addition to the measurement
information, using (6) a new assignment decision for the labels is made. This leads to an
iterative assignment of new labels.
If we write the contributions from kernel K1 , which depend only on x and do not change
with ? , as ch (j) = ?j ? yv(j) ? K1 (xh , xv(j) ) equation (6) becomes
yh (? ) = sgn
m
X
j=1
??j yv(j) K2 (yGh (? ? 1), yGv(j) ) + ch (j) + b , ?h.
(7)
Putting in the linear kernel from equation (5), we get
yh (? ) = sgn
m
X
j=1
?j yv(j)
?
q
X
?
|s|< c,s6=0
[yh+s (? ? 1) ? yv(j)+s ] + ch (j) + b , ?h. (8)
Interchanging the sums, using the definitions
? Pm
j=1 ?j yv(j) yv(j)+(k?h)
q
wh,k =
0
and
?h = ?
m
X
j=1
we obtain
yh (? ) = sgn
X
k
: k ? Gh
: k?
/ Gh
ch (j) + b ,
yk (? ? 1) ? wh,k ? ?h , ?h.
(9)
(10)
(11)
This corresponds to the equations describing the dynamics of a recurrent neural network
composed of McCulloch-Pitts neurons [9]. The condition for symmetric weights w h,k =
wk,h is equivalent to an inversion symmetry of the label configurations of the support vectors in the neighborhood topology, therefore the weights in equation (9) are not necessarily
symmetric. A suitable stopping criterion for the iteration is that the net reaches either a
fixed point yh (? ) = yh (? ? 1), ?h, or an attractor cycle yh (? ) = yh (?), ? < ? ? 1, ?h.
The network described by eq. (11) corresponds to a diluted network of ? binary neurons
with no self-interaction and asymmetric weights. One can see from eq.(9) that the network has only local connections, corresponding to the topographic neighborhood G h . The
measurement xh only influences the individual unit threshold ?h of the network, via the
weighted sum over all support vectors of the contributions from kernel K1 (eq. (10)). The
label configurations of the support vectors, on the other hand, are contained in the network
weights via eq.(9). The weights are multiplied by the hyper-parameter ?, which determines
how much the label configuration influences the class decision in comparison to the measurements. It has to be adjusted to yield optimal results for a class of data sets. For ? = 0
the TSVM becomes a conventional SVM.
2.5
Symmetrization of the weights
In order to ensure convergence, we suggest to use an inversion symmetric version K 2sym of
kernel K2 . For the pixel grid we can define the inversion operation as l + t ? l ? t, t ?
? . Taking the inverse of the vector yGl ,
N2 , ?l + t ? Gl , and denote the inverse of a by a
in which the set yGl is concatenated in an arbitrary but fixed order, leads to a reordering of
the components of the vector. The benefit from the chosen inversion symmetric kernel is
that the self consistency equations for the labels will turn out to be equivalent to a Hopfield
net, which has proven convergence properties. We define the new kernel as
K2sym (yGh , yGj ) =
1
(hyGh |yGj i + hyGh |?
yGj i).
q
(12)
Although only the second argument is inverted within the kernel, the value of this kernel
does not depend on the order of the arguments.
Proof It follows from the definition of the inversion operator and the dot product that
hyGh |yGj i = h?
yGh |?
yGj i = hyGj |yGh i = h?
yGj |?
yGh i and h?
yGh |yGj i = hyGh |?
yG j i =
hyGj |?
yGh i = h?
yGj |yGh i. Therefore,
K2sym (yGh , yGj )
=
=
1
1
(hyGh |yGj i + hyGh |?
yGj i) = (hyGj |yGh i + h?
yGj |yGh i)
q
q
1
(hyGj |yGh i + hyGj |?
yGh i) = K2sym (yGj , yGh )
q
Putting kernel (12) into eq.(7) and defining
? Pm
sym
j=1 ?j yv(j) (yv(j)+(k?h) + yv(j)?(k?h) )
q
wh,k =
0
: k ? Gh
: k?
/ Gh
.
(13)
we get
yh (? ) = sgn
X
k
sym
yk (? ? 1) ? wh,k
? ?h , ?h.
(14)
sym
Since the network weights wh,j
defined in eq.(13) are symmetric this corresponds to the
equation describing the dynamics during the retrieval phase of a Hopfield network [10].
Instead of taking the sum over all patterns, the sum is taken over all support vectors. The
weight between two neurons in the original Hopfield net corresponds to the correlation
between two components (over all fundamental patterns). In (13) the weight only depends
on the difference vector k-h between the two neurons on the 2D grid and is proportional to
the correlation (over all support vectors) between the label of a support vector and the label
in the distance k-h.
Table 1: Average misclassification rate R and the standard deviation of the mean ? at
optimal hyper-parameters C, S and ?.
algorithm
SVM
STSVM
TSVM
3
log2 C
4
4
2
log2 S
0.5
0.5
0.5
?
0
1.2
1.4
R[%]
2.23
1.96
1.86
?[%]
0.05
0.06
0.05
Experiments
We applied the above algorithms to the binary classification of pixel sites of cell images
from the CSSIP pap smear dataset. The goal was to assign the label +1 to the pixels belonging to the nucleus, and -1 to all others. The dataset contains three manual segmentations
of the nucleus? boundaries, from which we generated a ?ground truth? label for the area
of the nucleus using a majority voting. Only the first 300 images were used in the experiments. As a measurement vector we took a column-ordering of a 3x3 gray value patch
centered on a pixel site. In order to measure the classification performance for a noniid data set, we estimated the test error based on the collective classification of all pixels
in several randomly chosen test images. We compared three algorithms: A conventional
SVM, the ?TSVM? with the topographic kernel K2 from eq.(4) and the ?STSVM? with the
inversion symmetric topographic kernel K2sym from eq.(12). In the experiments we used
a label neighborhood of order 32, which corresponds to q = 100 neighbors. For kernel
K1 we used an RBF kernel K1 (x1 , x2 ) = exp(?kx1 ? x2 k2 /S 2 ) with hyper-parameter
S. Since the data set was very large, no cross-validation or resampling techniques were
required, and only a small subset of the available training data could be used for training.
We randomly sampled several disjoint training sets in order to improve the accuracy of
the error estimation. First, the hyper-parameters S and C (for TSVM and STSVM also ?)
were optimized via a grid search in parameter space. This was done by measuring the average test error over 20 test images and 5 training sets. Then, the test of the classifiers was
conducted at the in each case optimal hyper-parameters for 20 yet unused test images and
50 randomly sampled disjoint training sets. In all experiments using synchronous update
either a fixed point or an attractor cycle of length two was reached. The average number of
iterations needed was 12 (TSVM) and 13 (STSVM). Although the convergence properties
have only been formally proven for the symmetric weight STSVM, experimental evidence
suggests the same convergence properties for the TSVM. The results for synchronous update are shown in table 1 (results using asynchronous update differed only by 0.01%). The
performance of both topographic algorithms is superior to the conventional SVM, while
the TSVM performed slightly better than the STSVM. For the top-down method in [6] the
results were only qualitatively assessed by a human expert, not quantitatively compared to
a manual segmentation, therefore a direct comparison to our results was not possible. To
illustrate the role of the hyper-parameter ?, fig.1 shows 10 typical test images and their
segmentations achieved by an STSVM at different values of ? for fixed S and C. For
increasing ? the label images become less noisy, and at ? = 0.4 most artifacts have disappeared. This is caused by the increasing weight put on the label configuration via kernel
K2sym . Increasing ? even further will eventually lead to the appearance of spurious artifacts, as the influence of the label configuration will dominate the classification decision.
4
Conclusions
We have presented a classification method for a special case of non-iid data in which the
objects are linked by a regular grid structure. The proposed algorithm is composed of two
Figure 1: Final labels assigned by the STSVM at fixed hyper-parameters C = 2 6 , S = 22 .
(a) ? = 0, (b) ? = 0.1, (c) ? = 0.2, (d) ? = 0.3, (e) ? = 0.4.
components: The first part is a topographic kernel which integrates conventional feature
information and the information of the label configurations within a topographic neighborhood. The second part consists of a collective classification with recurrent neural network
dynamics which lets local label configurations converge to attractors determined by the
label configurations of the support vectors. For the asymmetric weight TSVM, the dimensionality of the problem is increased by the neighborhood size as compared to a conventional SVM (twice the neighborhood size for the symmetric weight STSVM). However,
the number of variables and constraints does not increase with the number of data points to
be labeled. Therefore, the TSVM and the STSVM can be applied to image segmentation
problems, where a large number of pixel labels have to be assigned simultaneously.
The algorithms were applied to the bottom-up cell nucleus segmentation in pap smear images needed for the detection of cervical cancer. The classification performance of the
TSVM and STSVM were compared to a conventional SVM, and it was shown that the inclusion of the topographic label configuration lead to a substantial decrease in the average
misclassification rate. The two topographic algorithms were much more resistant to noise
and smaller artifacts. A removal of artifacts which have similar size and the same measurement features as some of the nuclei cannot be achieved by a pure bottom-up method, as
this requires a priori model knowledge. In practice, the lower dimensional TSVM is to be
preferred over the STSVM, since it is faster and performed slightly better.
Acknowledgments
This work was funded by the BMBF (grant 01GQ0411). We thank Sepp Hochreiter for
useful discussions.
References
[1] J. Lafferty; A. McCallum; F. Pereira. Conditional random fields: Probabilistic models for
segmenting and labeling sequence data. In Proc. Int. Conf. on Machine Learning, 2001.
[2] S. Kumar; M. Hebert. Discriminative fields for modeling spatial dependencies in natural images. In Sebastian Thrun, Lawrence Saul, and Bernhard Scho? lkopf, editors, Advances in Neural
Information Processing Systems 16. MIT Press, Cambridge, MA, 2004.
[3] J. Neville and D. Jensen. Collective classification with relational dependency networks. In
Proc. 2nd Multi-Relational Data Mining Workshop, 9th ACM SIGKDD Intern. Conf. Knowledge
Discovery and Data Mining, 2003.
[4] B. Taskar, C. Guestrin, and D. Koller. Max-margin markov networks. In Sebastian Thrun,
Lawrence Saul, and Bernhard Scho? lkopf, editors, Advances in Neural Information Processing
Systems 16. MIT Press, Cambridge, MA, 2004.
[5] P. Bamford. Segmentation of Cell Images with Application to Cervical Cancer Screening. PhD
thesis, University of Queensland, 1999.
[6] P. Bamford and B. Lovell. Unsupervised cell nucleus segmentation with active contours. Signal
Processing Special Issue: Deformable Models and Techniques for Image and Signal Processing, 71(2):203?213, 1998.
[7] B. Sch?olkopf and A. Smola. Learning with Kernels. The MIT Press, 2002.
[8] V. Vapnik. Statistical Learning Theory. Springer, New York, 1998.
[9] W. McCulloch and W. Pitts. A logical calculus of the ideas immanent in nervous activity.
Bulletin of mathematical physics, 5:115?133, 1943.
[10] S. Haykin. Neural Networks. Macmillan College Publishing Company Inc., 1994.
| 2653 |@word version:1 inversion:6 polynomial:1 nd:1 calculus:1 queensland:2 configuration:18 contains:1 yet:2 written:1 shape:1 update:3 resampling:1 nervous:1 mccallum:1 haykin:1 location:1 mathematical:1 direct:1 become:1 consists:2 introduce:1 pairwise:1 multi:2 company:1 increasing:3 becomes:3 underlying:1 notation:1 mcculloch:2 lowest:1 voting:1 classifier:3 k2:10 unit:1 medical:1 grant:1 segmenting:1 engineering:1 local:7 xv:2 mohr:1 solely:1 might:1 twice:1 initialization:2 suggests:1 acknowledgment:1 yj:1 practice:2 block:1 x3:1 procedure:1 area:1 gabor:1 regular:3 suggest:1 get:2 cannot:1 unlabeled:1 operator:1 put:1 influence:3 equivalent:3 conventional:9 center:2 maximizing:1 go:4 sepp:1 rectangular:1 pure:1 rule:1 nuclear:1 dominate:1 s6:2 target:1 programming:2 us:1 recognition:1 asymmetric:2 labeled:3 bottom:4 role:1 taskar:1 electrical:1 hypertext:1 cycle:2 morphological:1 ordering:1 decrease:1 yk:2 substantial:1 dynamic:3 trained:2 depend:3 solving:1 hopfield:3 oby:1 labeling:2 klaus:1 hyper:8 neighborhood:13 whose:1 heuristic:1 solve:1 ability:1 gi:2 topographic:24 gp:1 noisy:2 final:2 sequence:2 net:3 took:1 propose:4 interaction:1 product:4 tu:2 neighboring:2 combining:1 relevant:2 achieve:1 kx1:1 deformable:1 olkopf:1 convergence:4 disappeared:1 object:6 diluted:1 illustrate:1 recurrent:3 measured:1 school:1 eq:9 c:2 filter:1 centered:1 human:2 sgn:6 require:1 assign:3 generalization:1 adjusted:1 correction:1 around:1 considered:1 ground:1 exp:1 lawrence:2 pitt:2 early:1 estimation:1 proc:2 integrates:1 label:52 symmetrization:1 successfully:1 weighted:1 mit:3 sensor:1 office:1 contrast:1 psychiatry:1 sigkdd:1 detect:1 stopping:1 initially:1 spurious:1 koller:1 going:1 germany:2 pixel:24 issue:1 classification:28 denoted:1 priori:1 spatial:1 special:3 field:4 manually:1 identical:1 unsupervised:1 interchanging:1 others:1 simplify:1 quantitatively:1 randomly:3 composed:3 simultaneously:1 individual:1 phase:3 ourselves:1 attractor:3 detection:2 screening:1 mining:2 johann:1 primal:1 watershed:1 necessary:1 instance:1 classify:2 column:1 increased:1 modeling:1 measuring:1 assignment:5 maximization:2 mac:1 deviation:1 subset:1 conducted:1 reported:1 dependency:6 referring:1 fundamental:1 probabilistic:2 physic:1 yg:1 thesis:1 central:1 receipt:1 possibly:1 conf:2 expert:1 de:2 accompanied:1 wk:1 includes:1 coefficient:1 int:1 inc:1 caused:1 depends:1 later:1 performed:2 observer:1 linked:1 reached:1 yv:10 tsvm:14 contribution:3 oi:1 accuracy:2 yield:1 weak:1 handwritten:1 lkopf:2 iid:4 reach:1 manual:2 sebastian:2 definition:2 associated:1 proof:1 hamming:1 sampled:2 dataset:3 wh:5 logical:1 knowledge:3 dimensionality:1 segmentation:14 higher:1 supervised:1 response:1 improved:1 evaluated:1 box:1 done:1 stage:1 smola:1 correlation:2 working:1 hand:1 nonlinear:1 artifact:4 gray:4 normalized:1 contain:1 assigned:3 symmetric:8 deal:1 adjacent:1 during:1 self:4 criterion:1 lovell:1 smear:3 gh:4 image:22 svc:1 scho:2 superior:1 physical:1 slight:1 measurement:17 cambridge:2 grid:9 pm:2 consistency:1 inclusion:1 centre:1 dot:3 funded:1 resistant:1 similarity:1 malignancy:1 charit:1 closest:1 binary:6 yi:4 inverted:2 guestrin:1 converge:1 signal:3 faster:1 characterized:2 cross:1 retrieval:1 divided:1 prediction:1 iteration:3 kernel:28 hochreiter:1 cell:7 achieved:2 addition:2 sch:1 posse:1 subject:1 lafferty:1 presence:1 unused:1 bernstein:2 identically:1 xj:2 restrict:1 topology:1 idea:2 shift:1 synchronous:2 papanicolaou:1 york:1 useful:1 johannes:1 cervical:4 canonical:1 neuroscience:2 stained:1 per:1 estimated:1 disjoint:2 write:1 putting:2 threshold:1 pap:3 sum:6 run:2 inverse:2 patch:2 decision:4 quadratic:2 activity:1 vectorial:1 constraint:3 x2:2 u1:1 argument:2 min:1 kumar:1 department:1 according:1 belonging:1 describes:1 slightly:2 smaller:1 psychotherapy:1 character:1 making:1 taken:1 equation:6 previously:1 slack:2 describing:2 turn:1 eventually:1 needed:2 available:3 operation:1 multiplied:1 apply:1 shortly:1 original:1 top:2 denotes:1 ensure:1 publishing:1 log2:2 giving:1 concatenated:2 k1:7 objective:1 dependence:1 ygj:18 obermayer:1 microscopic:1 gradient:2 distance:3 thank:1 mapped:1 berlin:7 separating:2 majority:1 thrun:2 topic:1 length:2 o1:1 index:3 relationship:1 neville:1 noniid:1 collective:7 unknown:3 neuron:4 markov:2 defining:1 relational:3 variability:1 extended:1 arbitrary:4 intensity:1 introduced:1 pair:1 required:1 connection:1 optimized:1 distinction:1 usually:2 pattern:2 max:1 suitable:1 misclassification:2 natural:3 difficulty:1 scheme:2 improve:3 movie:1 technology:1 text:1 review:1 discovery:1 removal:1 reordering:1 topography:2 proportional:1 proven:2 validation:1 clinic:1 nucleus:7 consistent:2 editor:2 bank:1 cancer:4 gl:1 asynchronous:1 sym:4 hebert:1 neighbor:8 saul:2 taking:2 bulletin:1 distributed:1 benefit:1 boundary:1 world:1 valid:1 contour:2 made:2 collection:1 qualitatively:1 preferred:1 bernhard:2 active:2 assumed:1 discriminative:3 xi:4 search:1 iterative:2 table:2 additionally:1 symmetry:1 depicting:1 hc:1 complex:1 necessarily:1 bamford:2 immanent:1 noise:1 n2:1 x1:1 site:9 fig:1 differed:1 bmbf:1 pereira:1 xh:8 yh:12 down:2 jensen:1 svm:10 evidence:1 exists:1 workshop:1 vapnik:1 texture:1 phd:1 margin:5 appearance:1 intern:1 visual:1 contained:1 macmillan:1 scalar:1 collectively:1 springer:1 ch:4 corresponds:5 truth:1 determines:1 acm:1 ma:2 conditional:2 goal:2 formulated:1 rbf:1 change:2 included:1 determined:3 typical:1 semantically:1 hyperplane:4 wt:1 called:3 experimental:1 subdivision:1 meaningful:1 formally:1 college:1 support:19 assessed:1 dissimilar:1 topographical:1 |
1,821 | 2,654 | Surface Reconstruction using
Learned Shape Models
Jan Erik Solem
School of Technology and Society
Malm?o University, Sweden
[email protected]
Fredrik Kahl
RSISE, Australian National University
ACT 0200, Australia
[email protected]
Abstract
We consider the problem of geometrical surface reconstruction from one
or several images using learned shape models. While humans can effortlessly retrieve 3D shape information, this inverse problem has turned out
to be difficult to perform automatically. We introduce a framework based
on level set surface reconstruction and shape models for achieving this
goal. Through this merging, we obtain an efficient and robust method for
reconstructing surfaces of an object category of interest.
The shape model includes surface cues such as point, curve and silhouette features. Based on ideas from Active Shape Models, we show how
both the geometry and the appearance of these features can be modelled
consistently in a multi-view context. The complete surface is obtained by
evolving a level set driven by a PDE, which tries to fit the surface to the
inferred 3D features. In addition, an a priori 3D surface model is used to
regularize the solution, in particular, where surface features are sparse.
Experiments are demonstrated on a database of real face images.
1 Introduction
Humans have a remarkable ability of perceiving 3D shape information - even from a single
photograph. Exactly how the human visual system works or how shape is represented is to
a large extent unknown. It is clear that the capabilities rely on a strong prior model and the
efficient use of different surface cues. The corresponding task of automatically recovering
a 3D surface model for a computer has turned out to be a challenging problem, even with
the addition of multiple images. In the present paper, we try to solve the problem in the
case of a single object category, more specifically, faces. At the heart of our approach lies
the combination of sophisticated surface reconstruction techniques and a strong statistical
model of surface features.
The first part concerns the statistical model of surface features for inferring 3D shape. The
features are primarily geometric ones, such as point, curve and silhouette features. Both the
geometric relations and their appearances (in terms of image intensities) are modelled. The
distributions are learned from real data. Also, a 3D model of the complete surface is used as
a weak regularizer where surface features are sparse. The motivations for introducing such
a model are several. We are interested in automatically recovering a surface model given
new image data of the object of interest. It is a hard problem to robustly extract curves and
apparent contours (i.e. silhouettes) without any a priori model. Moreover, many objects are
hard to reliably reconstruct due to specularities and illumination effects. By using distinct
geometric features and strong priors, we will still be able to obtain reliable results. Another
problem is textureless areas, and more generally, lack of information in the input. Our a
priori model will work as a domain-specific regularizer.
The second part of this work deals with fitting surfaces to points and curves in 3D space,
and at the same time, fitting the projections of surface contours to apparent contours in the
images. The approach taken here is a variational one - we define a functional which we will
try to minimize with respect to some parameters describing the geometry of the surfaces.
This variational problem leads to a surface evolution, driven by a PDE. The surface is
represented implicitly as the level set of a real-valued function [1].
1.1 Related Work
In the area of statistical shape models, our work is related to and inspired by Active Shape
Models (ASM) [2]. One distinction is that we will model both 2D and 3D data, while
ASM have mainly have been applied to 2D objects. In contrast to standard ASM, our
observations are from multiple views. In [3] a multi-view model is utilized, but no explicit
(or even consistent) 3D data is maintained within the model. In order to do inference on
our model, we have adopted the ideas of Probabilistic PCA (PPCA) [4].
In the seminal work [5], a complete 3D model is built from a database of 200 laser scans
of faces. The so-called morphable model can be fitted to an image with very impressive
results. The model itself is quite complex (approximately 70000 vertices), resulting in long
computation times. Although recent advances, their method still requires manual intervention [6]. A generic face model has also been used in [7] for computing regularized structure
and motion as well as in [8] based on silhouettes obtained from background subtraction.
In the area of using level set surface representations for fitting surfaces to 3D data, this paper
is related to the work in [9] where surfaces are fitted to points using a distance potential. In
[10] apparent contours were incorporated for level set based surface fitting. Surfaces can
also be estimated using 2D dense data, as in [11] based on photo-consistency. Shape priors
for level sets have previously been applied to segmentation, e.g. [12].
1.2 Contribution of the Paper
The main contribution of this paper is the approach itself - the combination of a state-of-theart surface reconstruction technique and the learned statistical model of surface features.
This merging results in a robust and efficient method for surface reconstruction without
any need for manual intervention. There is no need for an abundant number of images, in
fact, a single image is sufficient.
Another key contribution is the introduction of a multi-view feature model which is capable of representing both 2D and 3D data in a consistent manner. As the model is fully
probabilistic, the missing data problem can be handled in a natural way. By only incorporating distinct surface features, compared to a full morphable model, we not only gain
computational efficiency, but also robustness to specularities and other illumination effects.
This point is a also valid when compared to other surface reconstruction methods based on
image-correlation [11]. In the case of face modelling, this is known to cause problems due
to the complexity of the BRDF for human skin. The main contribution within the field of
level sets is in the incorporation of an a priori 3D model used for surface regularization.
2 Part I: A Learned Shape Model
In this section, we develop a statistical model which is needed in order to be able to automatically extract and compute 3D surface features from the input images. The output
from the model will serve as input to the level set reconstruction algorithm in order to get a
complete surface model. In addition, we will use an a priori 3D surface model as a (weak)
regularizer.
2.1 The Feature Model
Suppose we have a number of elements in a d-dimensional vector t, for example, a collection of 3D point coordinates. Suppose t can be related to some latent vector u of dimension
q where the relationship is linear:
t = W u + ?,
(1)
where W is a matrix of size d ? q and ? is d-vector allowing for non-zero mean. However, our measurements take place in the images, which is a non-linear function of the 3D
features according to the perspective camera model. Denote the projection function with
f : Rd ? Re , projecting all 3D features to 2D image features, for one or several images 1 .
Also, we need to change coordinate system of the 3D features to suit the actual projection function. Denote this mapping by T : Rd ? Rd . Thus, f (T (t)) will project all
normalized 3D data to all images.
Finally, a noise model needs to be specified. We assume that the image measurements
are independent and normally distributed, likewise, the latent variables are assumed to be
Gaussian with unit variance u ? N (0, I). Thus, in summary:
t2D = f (T (t)) + = f (T (W u + ?)) + ,
(2)
where ? N (0, ? 2 I) for some scalar ?.
Before the model can be used, its parameters need to be estimated from training data.
Given that it is a probabilistic model, this is best done with maximum likelihood (ML).
Suppose we are given n examples {t2D,i }ni=1 , the ML estimate for W and ? is obtained
by minimizing:
n
X
1
2
2
||(t2D,i ? f (Ti (W ui + ?))|| + ||ui || ,
(3)
?2
i=1
over all unknowns. The standard deviation ? is estimated a priori from the data. Once the
model parameters W and ? have been learned from examples, they are kept fix. In practice,
to minimize (3) we alternatively optimize over (W, ?) and the latent variables {u i }ni using
gradient descent. Initial estimates can be obtained by intersecting 3D structure from each
set of images and then applying standard PPCA algorithms for the linear part [4]. The
normalization Ti (?) is chosen such that each normalized 3D sample has zero mean and unit
variance.
A 3D point which is visible in m > 1 images will be represented in the vector t with its 3D
coordinates (X, Y, Z). For points visible in only one image, m = 1, no depth information
is available, and such points are represented similarly to apparent contour points. A curve
will be represented in the model by a number of points along the curve. In the training
of the model, it is important to parameterize each 3D curve such that each point on the
curve approximately corresponds to the same point on the corresponding curve in the other
examples. As for curves, we sample the apparent contours (in the images) using arc-length
parametrization. However, there is no 3D information available for the apparent contours
as they are view-dependent. A simple way is to treat contours points as 3D points with a
constant, approximate (but crude) depth estimate.
1
In the experiments, f (?) will model the projection of three calibrated perspective cameras.
2.2 The Grey-Level Model
The missing component in the model is the relationship between 2D image features and
the underlying grey-level (or color) values at these pixels. Again, we adopt a linear model
(PPCA). Using the same notation as in (1), but now with the subscript gl for grey-level, the
model can be written
tgl = Wgl ugl + ?gl + gl ,
where tgl is a vector containing the grey-level values of all the 2D image features and
gl is Gaussian noise in the measurements. In the training phase, each data sample of
grey-levels is normalized by subtracting the mean and scaling to unit variance. The MLestimate of Wgl and ?gl is computed with the EM-algorithm [4]. The complete statistical
two-layer model with one feature model and one grey-level model is very similar to the
concept of ASM [2]. In principle, the same techniques as used for ASM can be applied to
automatically compute the latent variables of the system, i.e. u and u gl .
2.3 The 3D Model
The two-layer feature model produces only a sparse set of features in 3D space. Even if
these cues are characteristic for a particular sample (or individual), it is often not enough
in order to infer a complete surface model, in particular, in regions where the features
are sparse. Therefore, we introduce a 3D surface model consisting of the complete mean
surface serving as domain-specific regularizer. The mean surface is obtained from laser
scans with the technique described in [13].
3 Part II: Surface Reconstruction
3.1 Level Set Formulation
Let x be a point in the open set ? ? R3 . The time dependent surface ?(t) is represented
implicitly as the zero level set of a function ?(x, t) : ? ? R+ ? R as
?(t) = {x ; ?(x, t) = 0} ,
(4)
where ? is defined such that ?(x, t) < 0 inside ? and ?(x, t) > 0 outside. Using the
definition above gives the outward unit normal n and the mean curvature ? as
n=
??
|??|
and ? = ? ?
??
.
|??|
(5)
One important, frequently used example is the signed distance function, where the requirement |??(x)| = 1 is imposed.
The zero set of ?(x, t) represents ?(t) at all times t. This means that ?(x(t), t) ? 0 for a
point on the curve x(t) ? ?(t). Differentiating with respect to t gives
?t + v ? ?? = 0 ? ?t + vn |??| = 0 ,
(6)
where v = dx(t)/dt and vn is the velocity normal to the surface. This PDE is solved
in order to move the surface according to some derived velocity v. For a more thorough
treatment of level set surfaces, see [1].
3.2 Surface Fitting to Points
In [9] surfaces are fitted to a point set S using the level set method. An initial surface is
deformed in the gradient direction of an energy functional which involves elastic energy
and potential energy. The energy is expressed using a distance potential as the surface
integral
Z
EP (?) =
d(x) d? ,
(7)
?
where ? is the surface, d? surface area and d(x) = dist(x, S) is the distance from the point
x to S. The gradient descent of (7) is, cf. [9],
?t = (?d(x) ? n + d(x)?) |??| ,
(8)
where n is the surface normal and ? the mean curvature.
This motion is known to interfere with surface regularization, since all surface points are
attracted to the 3D features. Therefore we cut the influence of the potential by setting
d(x) = min(d(x), dmax ).
3.3 Surface Fitting to Apparent Contours
Let ? be an apparent contour in an image parameterized as ?(s) : I ? R ? R 2 . The
back-projected cone, written in homogeneous coordinates,
C(s) = c + ?P + ?(s) ,
(9)
+
(where c is the camera center and P the pseudo-inverse of the 3 ? 4 camera matrix of
a perspective projection) should graze the surface at the location of the contour generator.
It is undesirable to attract the surface to the entire back-projected cone of the apparent
contour. The cone should only touch the surface along a curve - the so called contour
generator.
We propose to solve this in the following manner. For each point on the curve m = ?(s),
let x? denote the point closest to ? (If there are several, then choose the one with smallest
depth). The function ? is kept to be a signed distance function, i.e. |??| = 1. This means
that the point x? is easily found by checking the values of ? along the line of sight given
by (9). The set of these points, S 0 , will be a (possibly discontinuous) space curve ? ? (?, s).
This set is then added to the distance potential as
d(x) = min(d(x), d0 (x)) ,
(10)
0
0
where d (x) = dist(x, S ) is updated at appropriate intervals as the surface evolves.
3.4 Adding a 3D Shape Prior
Since the data, i.e. the points, curves and contours, are sparse it is customary to use a prior
for the regions where there is no information. Instead of the common choice of minimal
surface type models, we propose to use a learned shape model, as described in Section 2.3.
By first aligning the mean shape to the data, the deviation can be expressed similar to (7)
as
Z
EP rior (?) =
dP rior (x) d? ,
(11)
?
where dP rior (x) is the distance potential of the aligned mean shape.
3.5 The Combined Motion PDE
Adding all the components above gives a functional ET ot = EP +?EP rior , where ? ? R+
determines the weight of the prior shape. Combining the components (8), (10) and (11)
above leads to a PDE for the motion of the surface as
?t = [(?d(x) + ??dP rior (x)) ? n + (d(x) + ?dP rior (x))?]|??| .
(12)
This PDE is solved iteratively until a steady state is reached, which yields a (local) minimum of ET ot .
Figure 1: Extracted image features for one image triple and the reconstructed 3D surface.
4 Experiments
4.1 The Shape Model
All images were taken by a stereo setup with three (synchronized) cameras. The setup
was pre-calibrated for both extrinsic and intrinsic camera parameters. Example images are
given in Figures 1 and 2. In total a database of 28 image triplets were collected of faces
from different persons (23 males and 5 females of ages ranging from 7 to 65). 25 of these
were used for training and 3 for testing the two-layer feature model. The 3D mean surface
was computed from a database of 24 persons (different from the 28 persons above) using a
laser scanner as described in [13].
The 28 triplets in the training and test set were manually labelled - 36 points, 8 curves and
5 apparent contours were extracted for each person, see Figure 1. The two-layer model has
q = 12 elements in the latent variable u for the geometrical part and q gl = 15 elements
in ugl for the grey-level model. These numbers were found empirically in order to capture
most of the variations in the data. The model is able to adopt quite well to the test set. In
fact, even for one (frontal) input image of the test set, the model predicts the two profile
views remarkably well. As the pose over all faces is (essentially) constant for both test and
training images, the normalizing coordinate mapping T is restricted to scale and translation.
4.2 Surface Reconstruction
Once 3D data has been obtained by fitting the two-layer shape model to the input images,
surfaces are fitted to this data by solving the PDE (12). In the standard level set formulation,
the surface must be closed in the computational domain in order to make ? continuous. We
use the technique in [14] for initializing and evolving open implicit surfaces.
Surfaces were reconstructed for a number of persons, a selection is shown in Figures 1 and 2, where the zero set is visualized and triangulated using the marching cubes
algorithm . The value of ? was 0.3 and dmax was set from the maximum distance of the
feature points to the initial surface (4-5 voxels). For the reconstructions in Figure 2, the
mean and median distances (measured in voxel width) of the feature points to the reconstructed surface have been computed. See Figure 3 for a typical histogram and a table of
the results. Most points have sub-voxel distance to the surface, i.e. the deviation is of the
same order as the surface resolution. This shows that we are not fitting a surface to the
mean shape but that it really fits the feature data. The reader is encouraged to zoom in on
the reconstructions in Figure 2 to verify the quality.
5 Conclusions and Future Work
In this paper, a framework for reconstructing surface geometry based on learned shape
models and level sets has been presented. The approach relies solely on images as input
and the output of the system is a geometric 3D surface consistent with the input images. A
new regularization approach is introduced in the level set framework, where the distance to
an a priori 3D model is used instead of the common mean curvature regularization.
Figure 2: Input images and the reconstructed surface for three persons, two in the training
data and one (bottom) in the test data. For each person: One of the input images, triangulated surfaces and surfaces with texture. Note that the profiles above are different from the
ones in the input.
Histogram for distance to the surface for all 3D points
70
60
50
Person 1
Person 2
Person 3
40
30
20
10
0
0
mean
0.978
0.805
0.861
median
0.788
0.601
0.843
1
2
3
Distance in voxel width
Figure 3: Histogram of the deviations of the feature points from the surface for the second
person in Figure 2. The table displays mean and median deviations for all three persons.
Our current work focuses on incorporating a robust photo-consistency measure in the energy potential (7) to enable more detailed geometry. Also, the total number of faces is
quite small in the database and we will collect and label more images. Currently, only
images taken with the tri-stereo setup have been used with heads facing the middle camera (cf. Figure 1). Once the statistical model has been learned, it can be utilized for other
(nearby) poses as well, but to what extent is yet to be explored.
References
[1] J.A. Sethian. Level Set Methods and Fast Marching Methods Evolving Interfaces in Computational Geometry, Fluid Mechanics, Computer Vision, and Materials Science. Cambridge
University Press, 1999.
[2] T. F. Cootes and Taylor C. J. Active shape model search using local grey-level models: A
quantatitative evaluation. In British Machine Vision Conf., pages 639?648, 1993.
[3] T.F. Cootes, G.V. Wheeler, K.N. Walker, and C.J. Taylor. View-based active appearance models.
Image and Vision Computing, 20(9-10):657?664, 2002.
[4] M. E. Tipping and C. M. Bishop. Probabilistic principal component analysis. Phil. Trans. Royal
Soc. London B, 61(3):611?622, 1999.
[5] V. Blanz and T. Vetter. A morphable model for the synthesis of 3d faces. In SIGGRAPH, pages
187?194, 1999.
[6] S. Romdhani and T. Vetter. Efficient, robust and accurate fitting of a 3d morphable model. In
Int. Conf. Computer Vision, pages 59?66, Nice, France, 2003.
[7] P. Fua. Regularized bundle-adjustment to model heads from image sequences without calibration data. Int. J. Comput. Vision, 38(2):153?171, 2000.
[8] B. Moghaddam, J. Lee, H. Pfister, and R. Machiraju. Model-based 3d face capture with shapefrom-silhouettes. In IEEE International Workshop on Analysis and Modeling of Faces and
Gestures (AMFG), pages 20?27, 2003.
[9] H.K. Zhao, S. Osher, B. Merriman, and M. Kang. Implicit and non-parametric shape reconstruction from unorganized points using a variational level set method. In Computer Vision and
Image Understanding, pages 295?319, 2000.
[10] J.E. Solem and F. Kahl. Surface reconstruction from the projection of points, curves and contours. In 2nd Int. Symposium on 3D Data Processing, Visualization and Transmission, Thessaloniki, Greece, 2004.
[11] O. Faugeras and R. Keriven. Variational principles, surface evolution, PDEs, level set methods,
and the stereo problem. IEEE Transactions on Image Processing, 7(3):336?344, 1998.
[12] M. Rousson and N. Paragios. Shape priors for level set representations. In Proc. European
Conf. on Computer Vision, volume 2351 of Lecture Notes in Computer Science. Springer, 2002.
[13] K. Skoglund. Three-dimensional face modelling and analysis. Master?s thesis, Informatics and
Mathematical Modelling, Technical University of Denmark, DTU, 2003.
[14] J.E. Solem and A. Heyden. Reconstructing open surfaces from unorganized data points. In
International Conference on Computer Vision and Pattern Recognition, Washington DC, 2004.
| 2654 |@word deformed:1 middle:1 nd:1 open:3 grey:8 initial:3 kahl:2 current:1 yet:1 dx:1 written:2 attracted:1 must:1 visible:2 shape:25 cue:3 parametrization:1 math:1 location:1 mathematical:1 along:3 symposium:1 fitting:9 inside:1 manner:2 introduce:2 frequently:1 dist:2 multi:3 mechanic:1 inspired:1 automatically:5 actual:1 project:1 moreover:1 underlying:1 notation:1 what:1 pseudo:1 thorough:1 act:1 ti:2 exactly:1 normally:1 unit:4 intervention:2 before:1 local:2 treat:1 subscript:1 solely:1 approximately:2 signed:2 collect:1 challenging:1 camera:7 testing:1 practice:1 jan:1 wheeler:1 area:4 evolving:3 sethian:1 projection:6 pre:1 vetter:2 get:1 undesirable:1 selection:1 context:1 applying:1 seminal:1 influence:1 optimize:1 imposed:1 demonstrated:1 missing:2 center:1 phil:1 resolution:1 regularize:1 retrieve:1 coordinate:5 variation:1 updated:1 suppose:3 homogeneous:1 element:3 velocity:2 recognition:1 utilized:2 cut:1 predicts:1 database:5 ep:4 bottom:1 solved:2 capture:2 parameterize:1 initializing:1 region:2 complexity:1 ui:2 solving:1 serve:1 efficiency:1 easily:1 siggraph:1 represented:6 regularizer:4 laser:3 brdf:1 distinct:2 fast:1 london:1 outside:1 apparent:10 quite:3 faugeras:1 solve:2 valued:1 reconstruct:1 blanz:1 ability:1 asm:5 itself:2 mah:1 sequence:1 reconstruction:14 subtracting:1 propose:2 turned:2 aligned:1 combining:1 requirement:1 transmission:1 produce:1 object:5 develop:1 pose:2 measured:1 school:1 textureless:1 strong:3 soc:1 recovering:2 fredrik:2 involves:1 australian:1 synchronized:1 triangulated:2 direction:1 discontinuous:1 human:4 australia:1 enable:1 material:1 fix:1 really:1 scanner:1 effortlessly:1 normal:3 mapping:2 adopt:2 smallest:1 proc:1 label:1 currently:1 gaussian:2 sight:1 derived:1 focus:1 consistently:1 modelling:3 likelihood:1 mainly:1 contrast:1 inference:1 dependent:2 attract:1 entire:1 relation:1 france:1 interested:1 pixel:1 priori:7 t2d:3 cube:1 field:1 once:3 washington:1 manually:1 encouraged:1 represents:1 theart:1 future:1 primarily:1 national:1 zoom:1 individual:1 geometry:5 phase:1 consisting:1 suit:1 interest:2 evaluation:1 male:1 bundle:1 accurate:1 moghaddam:1 integral:1 capable:1 sweden:1 taylor:2 abundant:1 re:1 minimal:1 fitted:4 modeling:1 introducing:1 vertex:1 deviation:5 keriven:1 calibrated:2 combined:1 person:12 international:2 probabilistic:4 lee:1 informatics:1 synthesis:1 intersecting:1 again:1 thesis:1 containing:1 choose:1 possibly:1 conf:3 zhao:1 potential:7 includes:1 int:3 view:7 try:3 closed:1 reached:1 capability:1 contribution:4 minimize:2 ni:2 variance:3 characteristic:1 likewise:1 yield:1 modelled:2 weak:2 romdhani:1 manual:2 definition:1 energy:5 gain:1 ppca:3 treatment:1 color:1 jes:1 segmentation:1 greece:1 sophisticated:1 back:2 dt:1 tipping:1 fua:1 formulation:2 done:1 implicit:2 correlation:1 until:1 touch:1 lack:1 interfere:1 quality:1 effect:2 concept:1 verify:1 normalized:3 evolution:2 regularization:4 iteratively:1 deal:1 width:2 maintained:1 steady:1 complete:7 motion:4 interface:1 geometrical:2 ranging:1 image:41 variational:4 common:2 functional:3 empirically:1 volume:1 measurement:3 cambridge:1 rd:3 consistency:2 similarly:1 calibration:1 impressive:1 surface:83 morphable:4 aligning:1 curvature:3 closest:1 recent:1 female:1 perspective:3 driven:2 minimum:1 subtraction:1 ii:1 multiple:2 full:1 infer:1 d0:1 technical:1 gesture:1 malm:1 pde:7 long:1 essentially:1 vision:8 histogram:3 normalization:1 addition:3 remarkably:1 background:1 interval:1 median:3 walker:1 ot:2 tri:1 enough:1 fit:2 idea:2 pca:1 handled:1 stereo:3 cause:1 generally:1 se:2 clear:1 detailed:1 outward:1 visualized:1 category:2 estimated:3 extrinsic:1 serving:1 key:1 achieving:1 kept:2 cone:3 inverse:2 parameterized:1 master:1 cootes:2 place:1 tgl:2 reader:1 vn:2 scaling:1 layer:5 display:1 incorporation:1 nearby:1 min:2 according:2 combination:2 reconstructing:3 em:1 evolves:1 osher:1 projecting:1 restricted:1 heart:1 taken:3 visualization:1 previously:1 describing:1 r3:1 dmax:2 needed:1 photo:2 adopted:1 available:2 generic:1 appropriate:1 robustly:1 robustness:1 customary:1 cf:2 society:1 skin:1 move:1 added:1 parametric:1 gradient:3 dp:4 distance:13 extent:2 collected:1 denmark:1 erik:1 length:1 relationship:2 minimizing:1 difficult:1 setup:3 fluid:1 reliably:1 unknown:2 perform:1 allowing:1 observation:1 arc:1 descent:2 t:1 incorporated:1 head:2 dc:1 intensity:1 inferred:1 introduced:1 rsise:1 specified:1 learned:9 distinction:1 kang:1 unorganized:2 trans:1 able:3 pattern:1 built:1 reliable:1 royal:1 natural:1 rely:1 regularized:2 representing:1 technology:1 dtu:1 extract:2 prior:7 geometric:4 voxels:1 checking:1 nice:1 understanding:1 fully:1 lecture:1 facing:1 remarkable:1 generator:2 triple:1 age:1 heyden:1 sufficient:1 consistent:3 principle:2 translation:1 summary:1 gl:7 pdes:1 rior:6 face:12 differentiating:1 sparse:5 distributed:1 curve:17 dimension:1 depth:3 valid:1 contour:16 collection:1 projected:2 voxel:3 transaction:1 reconstructed:4 approximate:1 implicitly:2 silhouette:5 ml:2 active:4 assumed:1 alternatively:1 continuous:1 latent:5 search:1 triplet:2 table:2 robust:4 elastic:1 complex:1 european:1 domain:3 dense:1 main:2 motivation:1 noise:2 profile:2 sub:1 inferring:1 paragios:1 explicit:1 comput:1 lie:1 crude:1 british:1 specific:2 bishop:1 explored:1 concern:1 normalizing:1 incorporating:2 intrinsic:1 workshop:1 merging:2 adding:2 texture:1 illumination:2 marching:2 photograph:1 specularities:2 appearance:3 visual:1 expressed:2 adjustment:1 scalar:1 springer:1 corresponds:1 determines:1 relies:1 extracted:2 lth:1 goal:1 labelled:1 hard:2 change:1 specifically:1 perceiving:1 typical:1 principal:1 called:2 total:2 pfister:1 scan:2 frontal:1 |
1,822 | 2,655 | Maximum-Margin Matrix Factorization
Nathan Srebro
Dept. of Computer Science
University of Toronto
Toronto, ON, CANADA
[email protected]
Jason D. M. Rennie
Tommi S. Jaakkola
Computer Science and Artificial Intelligence Lab
Massachusetts Institute of Technology
Cambridge, MA, USA
jrennie,[email protected]
Abstract
We present a novel approach to collaborative prediction, using low-norm
instead of low-rank factorizations. The approach is inspired by, and has
strong connections to, large-margin linear discrimination. We show how
to learn low-norm factorizations by solving a semi-definite program, and
discuss generalization error bounds for them.
1
Introduction
Fitting a target matrix Y with a low-rank matrix X by minimizing the sum-squared error is
a common approach to modeling tabulated data, and can be done explicitly in terms of the
singular value decomposition of Y . It is often desirable, though, to minimize a different
loss function: loss corresponding to a specific probabilistic model (where X are the mean
parameters, as in pLSA [1], or the natural parameters [2]); or loss functions such as hinge
loss appropriate for binary or discrete ordinal data. Loss functions other than squared-error
yield non-convex optimization problems with multiple local minima. Even with a squarederror loss, when only some of the entries in Y are observed, as is the case for collaborative
filtering, local minima arise and SVD techniques are no longer applicable [3].
Low-rank approximations constrain the dimensionality of the factorization X = U V 0 .
Other constraints, such as sparsity and non-negativity [4], have also been suggested for
better capturing the structure in Y , and also lead to non-convex optimization problems.
In this paper we suggest regularizing the factorization by constraining the norm of U and
V ?constraints that arise naturally when matrix factorizations are viewed as feature learning for large-margin linear prediction (Section 2). Unlike low-rank factorizations, such
constraints lead to convex optimization problems that can be formulated as semi-definite
programs (Section 4). Throughout the paper, we focus on using low-norm factorizations
for ?collaborative prediction?: predicting unobserved entries of a target matrix Y , based on
a subset S of observed entries YS . In Section 5, we present generalization error bounds for
collaborative prediction using low-norm factorizations.
2
Matrix Factorization as Feature Learning
Using a low-rank model for collaborative prediction [5, 6, 3] is straightforward: A lowrank matrix X is sought that minimizes a loss versus the observed entries YS . Unobserved
entries in Y are predicted according to X. Matrices of rank at most k are those that can
be factored into X = U V 0 , U ? Rn?k , V ? Rm?k , and so seeking a low-rank matrix is
equivalent to seeking a low-dimensional factorization.
If one of the matrices, say U , is fixed, and only the other matrix V 0 needs to be learned, then
fitting each column of the target matrix Y is a separate linear prediction problem. Each row
of U functions as a ?feature vector?, and each column of V 0 is a linear predictor, predicting
the entries in the corresponding column of Y based on the ?features? in U .
In collaborative prediction, both U and V are unknown and need to be estimated. This can
be thought of as learning feature vectors (rows in U ) for each of the rows of Y , enabling
good linear prediction across all of the prediction problems (columns of Y ) concurrently,
each with a different linear predictor (columns of V 0 ). The features are learned without
any external information or constraints which is impossible for a single prediction task (we
would use the labels as features). The underlying assumption that enables us to do this in
a collaborative filtering situation is that the prediction tasks (columns of Y ) are related, in
that the same features can be used for all of them, though possibly in different ways.
Low-rank collaborative prediction corresponds to regularizing by limiting the dimensionality of the feature space?each column is a linear prediction problem in a low-dimensional
space. Instead, we suggest allowing an unbounded dimensionality for the feature space, and
regularizing by requiring a low-norm factorization, while predicting with large-margin.
Consider adding to the loss a penalty term which is the sum of squares of entries in U and
2
2
V , i.e. kU kFro + kV kFro (kkFro denotes the Frobenius norm). Each ?conditional? problem
(fitting U given V and vice versa) again decomposes into a collection of standard, this time
regularized, linear prediction problems. With an appropriate loss function, or constraints
on the observed entries, these correspond to large-margin linear discrimination problems.
For example, if we learn a binary observation matrix by minimizing a hinge loss plus such
a regularization term, each conditional problem decomposes into a collection of SVMs.
3
Maximum-Margin Matrix Factorizations
Matrices with a factorization X = U V 0 , where U and V have low Frobenius norm (recall
that the dimensionality of U and V is no longer bounded!), can be characterized in several
equivalent ways, and are known as low trace norm matrices:
Definition 1. The trace norm1 kXk? is the sum of the singular values of X.
2
2
Lemma 1. kXk? = minX=U V 0 kU kFro kV kFro = minX=U V 0 12 (kU kFro + kV kFro )
The characterization in terms of the singular value decomposition allows us to characterize
low trace norm matrices as the convex hull of bounded-norm rank-one matrices:
n
o
2
2
Lemma 2. {X | kXk? ? B} = conv uv 0 |u ? Rn , v ? Rm , |u| = |v| = B
In particular, the trace norm is a convex function, and the set of bounded trace norm matrices is a convex set. For convex loss functions, seeking a bounded trace norm matrix
minimizing the loss versus some target matrix is a convex optimization problem.
This contrasts sharply with minimizing loss over low-rank matrices?a non-convex problem. Although the sum-squared error versus a fully observed target matrix can be minimized efficiently using the SVD (despite the optimization problem being non-convex!),
minimizing other loss functions, or even minimizing a squared loss versus a partially observed matrix, is a difficult optimization problem with multiple local minima [3].
1
Also known as the nuclear norm and the Ky-Fan n-norm.
In fact, the trace norm has been suggested as a convex surrogate to the rank for various
rank-minimization problems [7]. Here, we justify the trace norm directly, both as a natural
extension of large-margin methods and by providing generalization error bounds.
To simplify presentation, we focus on binary labels, Y ? {?1}n?m . We consider hardmargin matrix factorization, where we seek a minimum trace norm matrix X that matches
the observed labels with a margin of one: Yia Xia ? 1 for all ia ? S. We also consider
soft-margin learning, where we minimize a trade-off between the trace norm of X and its
hinge-loss relative to YS :
X
minimize kXk? + c
max(0, 1 ? Yia Xia ).
(1)
ia?S
As in maximum-margin linear discrimination, there is an inverse dependence between the
norm and the margin. Fixing the margin and minimizing the trace norm is equivalent
to fixing the trace norm and maximizing the margin. As in large-margin discrimination
with certain infinite dimensional (e.g. radial)
pkernels, the data is always separable with
sufficiently high trace norm (a trace norm of n|S| is sufficient to attain a margin of one).
The max-norm variant Instead of constraining the norms of rows in U and V on average, we can constrain all rows of U and V to have small L2 norm, replacing the trace norm
with kXkmax = minX=U V 0 (maxi |Ui |)(maxa |Va |) where Ui , Va are rows of U, V . Lowmax-norm discrimination has a clean geometric interpretation. First, note that predicting
the target matrix with the signs of a rank-k matrix corresponds to mapping the ?items?
(columns) to points in Rk , and the ?users? (rows) to homogeneous hyperplanes, such that
each user?s hyperplane separates his positive items from his negative items. Hard-margin
low-max-norm prediction corresponds to mapping the users and items to points and hyperplanes in a high-dimensional unit sphere such that each user?s hyperplane separates his
positive and negative items with a large-margin (the margin being the inverse of the maxnorm).
4
Learning Maximum-Margin Matrix Factorizations
In this section we investigate the optimization problem of learning a MMMF, i.e. a low
norm factorization U V 0 , given a binary target matrix. Bounding the trace norm of U V 0 by
2
2
1
2 (kU kFro + kV kFro ), we can characterize the trace norm in terms of the trace of a positive
semi-definite matrix:
Lemma 3 ([7, Lemma 1]). For any X ? Rn?m
and t ? R: kXk? ? t iff there exists
A X
n?n
m?m
2
A?R
and B ? R
such that X 0 B < 0 and tr A + tr B ? 2t.
Proof. Note that for any matrix W , kW kFro = tr W W 0 . If XA0 X
< 0, we can write it as
B
2
2
1
0
U
0
0
a product [ V ] [ U V ]. We have X = U V and 2 (kU kFro + kV kFro ) = 21 (tr A + tr B) ? t,
0
establishing kXk? ? t. Conversely, if kXk? ? t we can write
it as X = U V with
0
0
0
U
U
X
tr U U + tr V V ? 2t and consider the p.s.d. matrix X 0 V V 0 .
Lemma 3 can be used in order to formulate minimizing the trace norm as a semi-definite
optimization problem (SDP). Soft-margin matrix factorization (1), can be written as:
X
yia Xia ? 1 ? ?ia
1
A X
min (tr A + tr B) + c
?ia s.t.
< 0,
?ia ? S (2)
0
X B
?ia ? 0
2
ia?S
2
A < 0 denotes A is positive semi-definite
Associating a dual variable Qia with each constraint on Xia , the dual of (2) is [8, Section
5.4.2]:
X
I
(?Q ? Y )
max
Qia s.t.
< 0, 0 ? Qia ? c
(3)
(?Q ? Y )0
I
ia?S
where Q ? Y denotes the sparse matrix (Q ? Y )ia = Qia Yia for ia ? S and zeros
elsewhere. The problem is strictly feasible, and there is no duality gap. The p.s.d. constraint
in the dual (3) is equivalent to bounding the spectral norm of Q ? Y , and the dual can also
be written as an optimization problem subject to a bound on the spectral norm, i.e. a bound
on the singular values of Q ? Y :
X
kQ ? Y k2 ? 1
max
Qia s.t.
(4)
0 ? Qia ? c ?ia ? S
ia?S
In typical collaborative prediction problems, we observe only a small fraction of the entries
in a large target matrix. Such a situation translates to a sparse dual semi-definite program,
with the number of variables equal to the number of observed entries. Large-scale SDP
solvers can take advantage of such sparsity.
The prediction matrix X ? minimizing (1) is part of the primal optimal solution of (2), and
can be extracted from it directly. Nevertheless, it is interesting to study how the optimal
prediction matrix X ? can be directly recovered from a dual optimal solution Q? alone.
Although unnecessary when relying on interior point methods used by most SDP solvers (as
these return a primal/dual optimal pair), this can enable us to use specialized optimization
methods, taking advantage of the simple structure of the dual.
Recovering X ? from Q? As for linear programming, recovering a primal optimal solution directly from a dual optimal solution is not always possible for SDPs. However, at
least for the hard-margin problem (no slack) this is possible, and we describe below how
an optimal prediction matrix X ? can be recovered from a dual optimal solution Q? by
calculating a singular value decomposition and solving linear equations.
Given a dual optimal Q? , consider its singular value decomposition Q? ? Y = U ?V 0 .
Recall that all singular values of Q? ?Y are bounded by one, and consider only the columns
? ? Rn?p of U and V? ? Rm?p of V with singular value one. It is possible to show [8,
U
Section 5.4.3], using complimentary slackness, that for some matrix R ? Rp?p , X ? =
? RR0 V? 0 is an optimal solution to the maximum margin matrix factorization problem (1).
U
Furthermore, p(p+1)
is bounded above by the number of non-zero Q?ia . When Q?ia > 0,
2
and assuming hard-margin constraints, i.e. no box constraints in the dual, complimentary
?
?i RR0 V? 0 = Yia , providing us with a linear equation on
slackness dictates that Xia
= U
a
p(p+1)
0
the
entries
in
the
symmetric
RR
. For hard-margin matrix factorization, we can
2
0
therefore recover the entries of RR by solving a system of linear equations, with a number
of variables bounded by the number of observed entries.
Recovering specific entries The approach described above requires solving a large system of linear equations (with as many variables as observations). Furthermore, especially
when the observations are very sparse (only a small fraction of the entries in the target
matrix are observed), the dual solution is much more compact then the prediction matrix:
the dual involves a single number for each observed entry. It might be desirable to avoid
storing the prediction matrix X ? explicitly, and calculate a desired entry Xi?0 a0 , or at least
its sign, directly from the dual optimal solution Q? .
Consider adding the constraint Xi0 a0 > 0 to the primal SDP (2). If there exists an optimal
solution X ? to the original SDP with Xi?0 a0 > 0, then this is also an optimal solution to
the modified SDP, with the same objective value. Otherwise, the optimal solution of the
modified SDP is not optimal for the original SDP, and the optimal value of the modified
SDP is higher (worse) than the optimal value of the original SDP.
Introducing the constraint Xi0 a0 > 0 to the primal SDP (2) corresponds to introducing a
new variable Qi0 a0 to the dual SDP (3), appearing in Q ? Y (with Yi0 a0 = 1) but not in the
objective. In this modified dual, the optimal solution Q? of the original dual would always
be feasible. But, if Xi?0 a0 < 0 in all primal optimal solutions, then the modified primal
SDP has a higher value, and so does the dual, and Q? is no longer optimal for the new dual.
By checking the optimality of Q? for the modified dual, e.g. by attempting to re-optimize
it, we can recover the sign of Xi?0 a0 .
We can repeat this test once with Yi0 a0 = 1 and once with Yi0 a0 = ?1, corresponding
to Xi0 a0 < 0. If Yi0 a0 Xi?0 a0 < 0 (in all optimal solutions), then the dual solution can be
improved by introducing Qi0 a0 with a sign of Yi0 a0 .
Predictions for new users So far, we assumed that learning is done on the known entries
in all rows. It is commonly desirable to predict entries in a new partially observed row of
Y (a new user), not included in the original training set. This essentially requires solving
a ?conditional? problem, where V is already known, and a new row of U is learned (the
predictor for the new user) based on a new partially observed row of X. Using maximummargin matrix factorization, this is a standard SVM problem.
Max-norm MMMF as a SDP The max-norm variant can also be written as a SDP, with
the primal and dual taking the forms:
min t + c
X
?ia s.t.
ia?S
max
X
ia?S
5
Qia s.t.
A
X0
?
(?Q ? Y )0
X
<0
B
Aii , Baa ? t ?i, a
yia Xia ? 1 ? ?ia
?ia ? S
?ia ? 0
(?Q ? Y )
<0
?
?, ? are diagonal
tr ? + tr ? = 1
0 ? Qia ? c ?ia ? S
(5)
(6)
Generalization Error Bounds for Low Norm Matrix Factorizations
Similarly to standard feature-based prediction approaches, collaborative prediction methods can also be analyzed in terms of their generalization ability: How confidently can we
predict entries of Y based on our error on the observed entries YS ? We present here generalization error bounds that holds for any target matrix Y , and for a random subset of
observations S, and bound the average error across all entries in terms of the observed
margin error3 . The central assumption, paralleling the i.i.d. source assumption for standard
feature-based prediction, is that the observed subset S is picked uniformly at random.
Theorem 4. For all target matrices Y ? {?1}n?m and sample sizes |S| > n log n, and
for a uniformly selected sample S of |S| entries in Y , with probability at least 1 ? ? over
3
The bounds presented here are special cases of bounds for general loss functions that we present
and prove elsewhere [8, Section 6.2]. To prove the bounds we bound the Rademacher complexity of
bounded trace norm and bounded max-norm matrices (i.e. balls w.r.t. these norms). The unit trace
norm ball is the convex hull of outer products of unit norm vectors. It is therefore enough to bound
the Rademacher complexity of such outer products, which boils down to analyzing the spectral norm
of random matrices. As a consequence of Grothendiek?s inequality, the unit max-norm ball is within
a factor of two of the convex hull of outer products of sign vectors. The Rademacher complexity of
such outer products can be bounded by considering their cardinality.
the sample selection, the following holds for all matrices X ? Rn?m and all ? > 0:
1
1
|{ia|Xia Yia ? 0}| <
|{ia ? S|Xia Yia ? ?}|+
nm
|S|
s
s
s
ln(1 + | log kXk? /?|)
kXk? ?
(n + m) ln n
ln(4/?)
4
ln m
K ?
+
+
|S|
|S|
2|S|
? nm
(7)
and
1
1
|{ia|Xia Yia ? 0}| <
|{ia ? S|Xia Yia ? ?}|+
nm
|S|
s
s
r
kXkmax n + m
ln(1 + | log kXk? /?|)
ln(4/?)
12
+
+
?
|S|
|S|
2|S|
(8)
Where K is a universal constant that does not depend on Y ,n,m,? or any other quantity.
To understand the scaling of these bounds, consider n ? m matrices X = U V 0 where
the norms of rows of U and V are bounded by?
r, i.e. matrices with kXkmax ? r2 . The
2
trace norm of such matrices is bounded by r / nm, and so the two bounds agree up to
log-factors?the cost of allowing the norm to be low on-average but not uniformly. Recall
that the conditional problem, where V is fixed and only U is learned, is a collection of
low-norm (large-margin) linear prediction problems. When the norms of rows in U and V
are bounded by r, aqsimilar generalization error bound on the conditional problem would
2
n
include the term r? |S|
, matching the bounds of Theorem 4 up to log-factors?learning
both U and V does not introduce significantly more error than learning just one of them.
Also
? of interest is the comparison with bounds for low-rank matrices, for which kXk? ?
?rank X kXkFro . In particular, for n ? m rank-k X with entries bounded by B, kXk? ?
knmB, and the second term in the right-hand side of (7) becomes:
s
B?
k(n + m) ln n
4
K
ln m
(9)
?
|S|
Although this is the best (up to log factors) that can be expected from scale-sensitive
bounds4 , taking a combinatorial approach, the dependence on the magnitude of the entries
in X (and the margin) can be avoided [9].
6
Implementation and Experiments
Ratings In many collaborative prediction tasks, the labels are not binary, but rather are
discrete ?ratings? in several ordered levels (e.g. one star through five stars). Separating R
levels by thresholds ?? = ?0 < ?1 < ? ? ? < ?R = ?, and generalizing hard-margin
constraints for binary labels, one can require ?Yia + 1 ? Xia ? ?Yia +1 ? 1. A soft-margin
version of these constraints, with slack variables for the two constraints on each observed
rating, corresponds to a generalization of the hinge loss which is a convex bound on the
zero/one level-agreement error (ZOE) [10]. To obtain a loss which is a convex bound on
the mean-absolute-error (MAE?the difference, in levels, between the predicted level and
the true level), we introduce R ? 1 slack variables for each observed rating?one for each
4
For general loss functions, bounds as in Theorem 4 depend only on the Lipschitz constant of
the loss, and (9) is the best (up to log factors) that can be achieved without explicitly bounding the
magnitude of the loss function.
of the R ? 1 constraints Xia ? ?r for r < Yia and Xia ? ?r for r ? Yia . Both of
these soft-margin problems (?immediate-threshold? and ?all-threshold?) can be formulated
as SDPs similar to (2)-(3). Furthermore, it is straightforward to learn also the thresholds
(they appear as variables in the primal, and correspond to constraints in the dual)?either
a single set of thresholds for the entire matrix, or a separate threshold vector for each row
of the matrix (each ?user?). Doing the latter allows users to ?use ratings differently? and
alleviates the need to normalize the data.
Experiments We conducted preliminary experiments on a subset of the 100K MovieLens
Dataset5 , consisting of the 100 users and 100 movies with the most ratings. We used CSDP
[11] to solve the resulting SDPs6 . The ratings are on a discrete scale of one through five,
and we experimented with both generalizations of the hinge loss above, allowing per-user
thresholds. We compared against WLRA and K-Medians (described in [12]) as ?Baseline?
learners. We randomly split the data into four sets. For each of the four possible test sets,
we used the remaining sets to calculate a 3-fold cross-validation (CV) error for each method
(WLRA, K-medians, trace norm and max-norm MMMF with immediate-threshold and allthreshold hinge loss) using a range of parameters (rank for WLRA, number of centers for
K-medians, slack cost for MMMF). For each of the four splits, we selected the two MMMF
learners with lowest CV ZOE and MAE and the two Baseline learners with lowest CV ZOE
and MAE, and measured their error on the held-out test data. Table 1 lists these CV and
test errors, and the average test error across all four test sets. On average and on three of
the four test sets, MMMF achieves lower MAE than the Baseline learners; on all four of
the test sets, MMMF achieves lower ZOE than the Baseline learners.
Test
Set
1
2
3
4
Method
WLRA rank 2
WLRA rank 2
WLRA rank 1
WLRA rank 2
Avg.
1
2
3
4
max-norm C=0.0012
trace norm C=0.24
max-norm C=0.0012
max-norm C=0.0012
Avg.
ZOE
CV
Test
0.547 0.575
0.550 0.562
0.562 0.543
0.557 0.553
0.558
0.543 0.562
0.550 0.552
0.551 0.527
0.544 0.550
0.548
Method
K-Medians K=2
K-Medians K=2
K-Medians K=2
K-Medians K=2
max-norm C=0.0012
max-norm C=0.0011
max-norm C=0.0012
max-norm C=0.0012
MAE
CV
Test
0.678 0.691
0.686 0.681
0.700 0.681
0.685 0.696
0.687
0.669 0.677
0.675 0.683
0.668 0.646
0.667 0.686
0.673
Table 1: Baseline (top) and MMMF (bottom) methods and parameters that achieved the
lowest cross validation error (on the training data) for each train/test split, and the error for
this predictor on the test data. All listed MMMF learners use the ?all-threshold? objective.
7
Discussion
Learning maximum-margin matrix factorizations requires solving a sparse semi-definite
program. We experimented with generic SDP solvers, and were able to learn with up to
tens of thousands of labels. We propose that just as generic QP solvers do not perform
well on SVM problems, special purpose techniques, taking advantage of the very simple
structure of the dual (3), are necessary in order to solve large-scale MMMF problems.
SDPs were recently suggested for a related, but different, problem: learning the features
5
http://www.cs.umn.edu/Research/GroupLens/
Solving with immediate-threshold loss took about 30 minutes on a 3.06GHz Intel Xeon.
Solving with all-threshold loss took eight to nine hours. The MATLAB code is available at
www.ai.mit.edu/?nati/mmmf
6
(or equivalently, kernel) that are best for a single prediction task [13]. This task is hopeless
if the features are completely unconstrained, as they are in our formulation. Lanckriet et al
suggest constraining the allowed features, e.g. to a linear combination of a few ?base feature spaces? (or base kernels), which represent the external information necessary to solve
a single prediction problem. It is possible to combine the two approaches, seeking constrained features for multiple related prediction problems, as a way of combining external
information (e.g. details of users and of items) and collaborative information.
An alternate method for introducing external information into our formulation is by adding
to U and/or V additional fixed (non-learned) columns representing the external features.
This method degenerates to standard SVM learning when Y is a vector rather than a matrix.
An important limitation of the approach we have described, is that observed entries are
assumed to be uniformly sampled. This is made explicit in the generalization error bounds.
Such an assumption is typically unrealistic, as, e.g., users tend to rate items they like. At
an extreme, it is often desirable to make predictions based only on positive samples. Even
in such situations, it is still possible to learn a low-norm factorization, by using appropriate
loss functions, e.g. derived from probabilistic models incorporating the observation process. However, obtaining generalization error bounds in this case is much harder. Simply
allowing an arbitrary sampling distribution and calculating the expected loss based on this
distribution (which is not possible with the trace norm, but is possible with the max-norm
[8]) is not satisfying, as this would guarantee low error on items the user is likely to want
anyway, but not on items we predict he would like.
Acknowledgments We would like to thank Sam Roweis for pointing out [7].
References
[1] T. Hofmann. Unsupervised learning by probabilistic latent semantic analysis. Machine Learning Journal, 42(1):177?196, 2001.
[2] M. Collins, S. Dasgupta, and R. Schapire. A generalization of principal component analysis to
the exponential family. In Advances in Neural Information Processing Systems 14, 2002.
[3] Nathan Srebro and Tommi Jaakkola. Weighted low rank approximation. In 20th International
Conference on Machine Learning, 2003.
[4] D.D. Lee and H.S. Seung. Learning the parts of objects by non-negative matrix factorization.
Nature, 401:788?791, 1999.
[5] T. Hofmann. Latent semantic models for collaborative filtering. ACM Trans. Inf. Syst.,
22(1):89?115, 2004.
[6] Benjamin Marlin. Modeling user rating profiles for collaborative filtering. In Advances in
Neural Information Processing Systems, volume 16, 2004.
[7] Maryam Fazel, Haitham Hindi, and Stephen P. Boyd. A rank minimization heuristic with application to minimum order system approximation. In Proceedings American Control Conference,
volume 6, 2001.
[8] Nathan Srebro. Learning with Matrix Factorization. PhD thesis, Massachusetts Institute of
Technology, 2004.
[9] N. Srebro, N. Alon, and T. Jaakkola. Generalization error bounds for collaborative prediction
with low-rank matrices. In Advances In Neural Information Processing Systems 17, 2005.
[10] Amnon Shashua and Anat Levin. Ranking with large margin principle: Two approaches. In
Advances in Neural Information Proceedings Systems, volume 15, 2003.
[11] B. Borchers. CSDP, a C library for semidefinite programming. Optimization Methods and
Software, 11(1):613?623, 1999.
[12] B. Marlin. Collaborative filtering: A machine learning perspective. Master?s thesis, University
of Toronto, 2004.
[13] G. Lanckriet, N. Cristianini, P. Bartlett, L. El Ghaoui, and M. Jordan. Learning the kernel matrix
with semidefinite programming. Journal of Machine Learning Research, 5:27?72, 2004.
| 2655 |@word version:1 norm:65 yi0:5 plsa:1 seek:1 decomposition:4 tr:11 harder:1 recovered:2 written:3 hofmann:2 enables:1 discrimination:5 alone:1 intelligence:1 selected:2 item:9 characterization:1 toronto:4 hyperplanes:2 five:2 unbounded:1 prove:2 fitting:3 combine:1 introduce:2 x0:1 expected:2 sdp:16 inspired:1 relying:1 solver:4 qia:8 conv:1 considering:1 cardinality:1 underlying:1 bounded:14 becomes:1 lowest:3 complimentary:2 minimizes:1 maxa:1 unobserved:2 marlin:2 guarantee:1 rm:3 k2:1 control:1 unit:4 appear:1 positive:5 local:3 consequence:1 despite:1 analyzing:1 establishing:1 might:1 plus:1 conversely:1 factorization:26 range:1 fazel:1 acknowledgment:1 definite:7 universal:1 thought:1 attain:1 dictate:1 matching:1 significantly:1 radial:1 boyd:1 suggest:3 interior:1 selection:1 impossible:1 optimize:1 equivalent:4 www:2 center:1 maximizing:1 straightforward:2 convex:15 formulate:1 factored:1 nuclear:1 his:3 anyway:1 limiting:1 target:11 user:15 paralleling:1 programming:3 homogeneous:1 agreement:1 lanckriet:2 satisfying:1 observed:19 bottom:1 calculate:2 thousand:1 trade:1 benjamin:1 ui:2 complexity:3 seung:1 cristianini:1 depend:2 solving:8 learner:6 completely:1 aii:1 differently:1 various:1 train:1 describe:1 borchers:1 artificial:1 heuristic:1 solve:3 rennie:1 say:1 otherwise:1 ability:1 advantage:3 rr:2 took:2 propose:1 maryam:1 product:5 combining:1 iff:1 alleviates:1 degenerate:1 roweis:1 frobenius:2 kv:5 normalize:1 ky:1 rademacher:3 object:1 alon:1 fixing:2 measured:1 lowrank:1 strong:1 recovering:3 c:2 predicted:2 involves:1 tommi:3 hull:3 qi0:2 enable:1 require:1 generalization:13 preliminary:1 extension:1 strictly:1 hold:2 sufficiently:1 mapping:2 predict:3 pointing:1 sought:1 achieves:2 purpose:1 applicable:1 label:6 combinatorial:1 grouplens:1 sensitive:1 vice:1 weighted:1 minimization:2 mit:2 concurrently:1 always:3 modified:6 rather:2 avoid:1 jaakkola:3 derived:1 focus:2 rank:24 contrast:1 baseline:5 el:1 entire:1 typically:1 a0:15 dual:25 constrained:1 special:2 equal:1 once:2 sampling:1 kw:1 unsupervised:1 minimized:1 simplify:1 few:1 randomly:1 consisting:1 interest:1 investigate:1 umn:1 analyzed:1 extreme:1 semidefinite:2 primal:9 held:1 necessary:2 desired:1 re:1 column:10 modeling:2 soft:4 xeon:1 norm1:1 cost:2 introducing:4 entry:26 subset:4 predictor:4 kq:1 csdp:2 levin:1 conducted:1 characterize:2 mmmf:11 international:1 csail:1 probabilistic:3 off:1 lee:1 thesis:2 squared:4 again:1 central:1 nm:4 possibly:1 worse:1 external:5 yia:14 american:1 return:1 syst:1 star:2 explicitly:3 ranking:1 jason:1 lab:1 picked:1 doing:1 shashua:1 xa0:1 recover:2 collaborative:16 minimize:3 square:1 efficiently:1 yield:1 correspond:2 sdps:3 definition:1 against:1 naturally:1 proof:1 boil:1 sampled:1 massachusetts:2 recall:3 dimensionality:4 wlra:7 higher:2 improved:1 formulation:2 done:2 though:2 box:1 furthermore:3 just:2 hand:1 replacing:1 slackness:2 usa:1 requiring:1 true:1 regularization:1 symmetric:1 semantic:2 novel:1 recently:1 common:1 specialized:1 qp:1 volume:3 interpretation:1 xi0:3 mae:5 he:1 cambridge:1 versa:1 rr0:2 cv:6 ai:1 uv:1 unconstrained:1 similarly:1 jrennie:1 longer:3 base:2 perspective:1 inf:1 certain:1 inequality:1 binary:6 minimum:5 additional:1 stephen:1 semi:7 multiple:3 desirable:4 maximummargin:1 match:1 characterized:1 cross:2 sphere:1 y:4 va:2 prediction:32 variant:2 baa:1 essentially:1 kernel:3 represent:1 achieved:2 want:1 singular:8 source:1 median:7 unlike:1 subject:1 tend:1 jordan:1 constraining:3 split:3 enough:1 associating:1 translates:1 amnon:1 bartlett:1 penalty:1 tabulated:1 nine:1 matlab:1 listed:1 ten:1 svms:1 http:1 schapire:1 sign:5 estimated:1 per:1 discrete:3 write:2 dasgupta:1 four:6 nevertheless:1 threshold:11 clean:1 fraction:2 sum:4 inverse:2 master:1 throughout:1 family:1 scaling:1 capturing:1 bound:24 fan:1 fold:1 constraint:16 sharply:1 constrain:2 software:1 nathan:3 min:2 optimality:1 attempting:1 separable:1 according:1 alternate:1 ball:3 combination:1 across:3 sam:1 ghaoui:1 ln:8 equation:4 agree:1 discus:1 slack:4 ordinal:1 available:1 eight:1 observe:1 appropriate:3 spectral:3 generic:2 appearing:1 rp:1 original:5 denotes:3 remaining:1 include:1 dataset5:1 top:1 hinge:6 calculating:2 especially:1 seeking:4 objective:3 already:1 quantity:1 zoe:5 dependence:2 diagonal:1 surrogate:1 minx:3 separate:4 thank:1 separating:1 outer:4 assuming:1 code:1 providing:2 minimizing:9 equivalently:1 difficult:1 trace:25 negative:3 implementation:1 unknown:1 perform:1 allowing:4 observation:5 enabling:1 immediate:3 situation:3 rn:5 arbitrary:1 canada:1 rating:8 squarederror:1 pair:1 connection:1 learned:5 hour:1 trans:1 able:1 suggested:3 below:1 sparsity:2 confidently:1 program:4 max:19 ia:25 unrealistic:1 natural:2 regularized:1 predicting:4 hindi:1 representing:1 movie:1 technology:2 library:1 negativity:1 geometric:1 l2:1 nati:2 checking:1 relative:1 loss:28 fully:1 interesting:1 limitation:1 filtering:5 srebro:4 versus:4 validation:2 sufficient:1 principle:1 haitham:1 storing:1 row:14 elsewhere:2 hopeless:1 repeat:1 side:1 understand:1 institute:2 taking:4 absolute:1 sparse:4 ghz:1 xia:13 collection:3 commonly:1 avg:2 avoided:1 made:1 far:1 compact:1 maxnorm:1 unnecessary:1 assumed:2 xi:5 latent:2 decomposes:2 table:2 learn:5 ku:5 nature:1 obtaining:1 bounding:3 arise:2 profile:1 allowed:1 kxkmax:3 intel:1 explicit:1 exponential:1 anat:1 rk:1 theorem:3 down:1 minute:1 specific:2 maxi:1 r2:1 experimented:2 svm:3 list:1 exists:2 incorporating:1 adding:3 phd:1 magnitude:2 margin:32 gap:1 generalizing:1 simply:1 likely:1 kxk:12 ordered:1 partially:3 corresponds:5 extracted:1 ma:1 acm:1 conditional:5 viewed:1 formulated:2 presentation:1 lipschitz:1 feasible:2 hard:5 included:1 infinite:1 typical:1 uniformly:4 movielens:1 justify:1 hyperplane:2 lemma:5 principal:1 duality:1 svd:2 latter:1 collins:1 dept:1 regularizing:3 |
1,823 | 2,656 | Outlier Detection with One-class Kernel Fisher
Discriminants
Volker Roth
ETH Zurich, Institute of Computational Science
Hirschengraben 84, CH-8092 Zurich
[email protected]
Abstract
The problem of detecting ?atypical objects? or ?outliers? is one of the
classical topics in (robust) statistics. Recently, it has been proposed to
address this problem by means of one-class SVM classifiers. The main
conceptual shortcoming of most one-class approaches, however, is that in
a strict sense they are unable to detect outliers, since the expected fraction
of outliers has to be specified in advance. The method presented in this
paper overcomes this problem by relating kernelized one-class classification to Gaussian density estimation in the induced feature space. Having
established this relation, it is possible to identify ?atypical objects? by
quantifying their deviations from the Gaussian model. For RBF kernels
it is shown that the Gaussian model is ?rich enough? in the sense that it
asymptotically provides an unbiased estimator for the true density. In order to overcome the inherent model selection problem, a cross-validated
likelihood criterion for selecting all free model parameters is applied.
1
Introduction
A one-class-classifier attempts to find a separating boundary between a data set and the
rest of the feature space. A natural application of such a classifier is estimating a contour
line of the underlying data density for a certain quantile value. Such contour lines may
be used to separate ?typical? objects from ?atypical? ones. Objects that look ?sufficiently
atypical? are often considered to be outliers, for which one rejects the hypothesis that they
come from the same distribution as the majority of the objects. Thus, a useful application
scenario would be to find a boundary which separates the jointly distributed objects from
the outliers. Finding such a boundary defines a classification problem in which, however,
usually only sufficiently many labeled samples from one class are available. Usually no
labeled samples from the outlier class are available at all, and it is even unknown if there
are any outliers present.
It is interesting to notice that the approach of directly estimating a boundary, as opposed to
first estimating the whole density, follows one of the main ideas in learning theory which
states that one should avoid solving a too hard intermediate problem. While this line of reasoning seems to be appealing from a theoretical point of view, it leads to a severe problem
in practical applications: when it comes to detecting outliers, the restriction to estimating
only a boundary makes it impossible to derive a formal characterization of outliers without prior assumptions on the expected fraction of outliers or even on their distribution. In
practice, however, any such prior assumptions can hardly be justified. The fundamental
problem of the one-class approach lies in the fact that outlier detection is a (partially) unsupervised task which has been ?squeezed? into a classification framework. The missing part
of information has been shifted to prior assumptions which can probably only be justified,
if the solution of the original problem was known in advance.
This paper aims at overcoming this problem by linking kernel-based one-class classifiers
to Gaussian density estimation in the induced feature space. Objects which have an ?unexpected? high Mahalanobis distance to the sample mean are considered as ?atypical objects?
or outliers. A particular Mahalanobis distance is considered to be unexpected, if it is very
unlikely to observe an object that far away from the mean vector in a random sample of
a certain size. We will formalize this concept in section 3 by way of fitting linear models
in quantile-quantile plots. The main technical ingredient of our method is the one-class
kernel Fisher discriminant classifier (OC-KFD), for which the relation to Gaussian density
estimation is shown. From the classification side, the OC-KFD-based model inherits the
simple complexity control mechanism by using regularization techniques. The explicit relation to Gaussian density estimation, on the other hand, makes it possible to formalize the
notion of atypical objects by observing deviations from the Gaussian model. It is clear that
these deviations will heavily depend on the chosen model parameters. In order to derive
an objective characterization of atypical objects it is, thus, necessary to select a suitable
model in advance. This model-selection problem is overcome by using a likelihood-based
cross-validation framework for inferring the free parameters.
2
Gaussian density estimation and one-class LDA
Let X denote the n ? d data matrix which contains the n input vectors xi ? Rd as rows. It
has been proposed to estimate a one-class decision boundary by separating the dataset from
the origin [12], which effectively coincides with replicating all xi with opposite sign and
separating X and ?X. Typically, a ?-SVM classifier with RBF kernel function is used.
The parameter ? bounds the expected number of outliers and must be selected a priori. The
method proposed here follows the same idea of separating the data from their negatively
replicated counterparts. Instead of a SVM, however, a Kernel Fisher Discriminant (KFD)
classifier is used [7, 10]. The latter has the advantage that is is closely related to Gaussian
density estimation in the induced feature space. By making this relation explicit, outliers
can be identified without specifying the expected fraction of outliers in advance. We start
with a linear discriminant analysis (LDA) model, and then kernels will be introduced.
Let X 0 = (X, ?X)> denote the augmented (2n ? d) data matrix which also contains
the negative
P samples ?xi . Without loss of generality we assume that the sample mean
?+ ?
i xi > 0, so that the sample means of the positive data and the negative data
differ: ?+ 6= ?? . Let us now assume that our data are realizations of a normally distributed
random variable in d dimensions: X ? Nd (?, ?). Denoting by X c the centered data
? ? W = (1/n)X c> X c .
matrix, the estimator for ? takes the form ?
>
The LDA solution ? ? maximizes the between-class scatter ? >
? B? ? with B = ?+ ?+ +
>
>
?? ?? under the constraint on the within-class scatter ? ? W ? ? = 1. Note that in our
special case with X 0 = (X, ?X)> the usual pooled within-class matrix W simply reduces
to the above defined W = (1/n)X c> X c . Denoting by y 0 = (2, . . . , 2, ?2, . . . , ?2)> a
2n-indicator vector for class membership in class ?+? or ???, it is well-known (see e.g. [1])
that the LDA solution (up to a scaling factor) can be found by minimizing a least-squares
? = arg min? ky 0 ?X 0 ?k2 . In [3] a slightly more general form of the problem
functional: ?
is described where the above functional is minimized under a constrained on ?, which in
the simplest case amounts to adding a term ?? > ? to the functional. Such a ridge regression
model assumes a penalized total covariance of the form T = (1/(2n)) ? X 0> X 0 + ?I =
?
(1/n) ? X > X + ?I. Defining an n-vector of ones y = (1, . . . , 1)> , the solution vector ?
reads
? = (X 0> X 0 + ?I)?1 X 0> y 0 = (X > X + ?I)?1 X > y.
?
(1)
2
According to [3], an appropriate scaling factor is defied in terms of the quantity s =
? which leads us to the correctly scaled LDA vector ? ? =
(1/n) ? y > y? = (1/n) ? y > X ?,
?1
2 ?1/2 ?
s (1 ? s )
? that fulfills the normalization condition ? >
? W ? ? = 1.
One further derives from [3] that the mean vector of X, projected onto the 1-dimensional
LDA-subspace has the coordinate value m+ = s(1 ? s2 )?1/2 , and that the Mahalanobis
distance from a vector x to the sample mean ?+ is the sum of the squared Euclidean
distance in the projected space and an orthogonal distance term:
>
2
2
2
> ?1
D(x, ?+ ) = (? >
x.
? x ? m+ ) + D? with D? = ?(1 ? s )(? ? x) + x T
(2)
Note that it is the term D? which makes the density estimation model essentially different
from OC-classification: while the latter considers only distances in the direction of the
projection vector ?, the true density model additionally takes into account the distances in
the orthogonal subspace.
Since the assumption X ? Nd (?, ?) is very restrictive, we propose to relax it by assuming
that we have found a suitable transformation of our input data ? : Rd 7? Rp , x 7? ?(x),
such that the transformed data are Gaussian in p dimensions. If the transformation is carried
out implicitly by introducing a Mercer kernel k(xi , xj ), we arrive at an equivalent problem
in terms of the kernel matrix K = ??> and the expansion coefficients ?:
?
? = (K + ?I)?1 y.
(3)
From [11] it follows that the mapped vectors can be represented in Rn as ?(x) =
K ?1/2 k(x), where k(x) denotes the kernel vector k(x) = (k(x, x1 ), . . . , k(x, xn ))> .
Finally we derive the following form of the Mahalanobis distances which again consist of
the Euclidean distance in the classification subspace plus an orthogonal term:
2
2
>
2
D(x, ?+ ) = (?>
? k(x) ? m+ ) ? (1 ? s )(?? k(x)) + n?(x),
(4)
where ?(x) = ?> (x)(?> ? + ?I)?1 ?(x) = k> (x)(K + ?I)?1 K ?1 k(x),
m+ = s(1 ? s2 )?1/2 , s2 = (1/n) ? y > y? = (1/n) ? y > K ?,
? and ?? = s?1 (1 ? s2 )?1/2 ?.
?
Equation (4) establishes the desired link between OC-KFD and Gaussian density estimation, since for our outlier detection mechanism only Mahalanobis distances are needed.
While it seems to be rather complicated to estimate a density by the above procedure, the
main benefit over directly estimating the mean and the covariance lies in the inherent complexity regulation properties of ridge regression. Such a complexity control mechanism is
of particular importance in highly nonlinear kernel models. Moreover, for ridge-regression
models it is possible to analytically calculate the effective degrees of freedom, a quantity
that will be of particular interest when it comes to detecting outliers.
3
Detecting outliers
Let us assume that the model is completely specified, i.e. both the kernel function k(?, ?)
and the regularization parameter ? are fixed. The central lemma that helps us to detect
outliers can be found in most statistical textbooks:
Lemma 1. Let X be a Gaussian random variable X ? Nd (?, ?). Then ? ? (X ?
?)> ??1 (X ? ?) follows a chi-square (?2 ) distribution on d degrees of freedom.
For the penalized regression models, it might be more appropriate to use the effective degrees of freedom df instead of d in the above lemma. In the case of one-class LDA with
ridge penalties we can easily estimate it as df = trace(X(X > X + ?I)?1 X > ), [8], which
for a kernel model translates into df = trace(K(K + ?I)?1 ). The intuitive interpretation
of the quantity df is the following: denoting by V the matrix of eigenvectors of K and by
{?i }ni=1 the corresponding eigenvalues, the fitted values y? read
y? = V diag {?i = ?i /(?i + ?)} V > y.
(5)
It follows that compared to the unpenalized case, where all eigenvectors v i are constantly
weighted by 1, the contribution of the i-th eigenvector v i is down-weighted by a factor
?i /1 = ?i . If the ordered eigenvalues decrease rapidly, however, the values ?i are either
close to zero or close to one, and df determines the number of terms that are ?essentially
different? from zero. The same is true for the orthogonal distance term in eq. (4): note that
?(x) = k> (x)(K + ?I)?1 K ?1 k(x) = k> V diag ?i0 = ((?i + ?)?i )?1 V > k(x). (6)
Compared to the unpenalized case (the contribution of v i is weighted by ??2
i ), the contribution of v i is down-weighted by the same factor ?i0 /??2
= ?i .
i
From lemma 1 we conclude that if the data are well described by a Gaussian model in
the kernel feature space, the observed Mahalanobis distances should look like a sample
from a ?2 -distribution with df degrees of freedom. A graphical way to test this hypothesis
is to plot the observed quantiles against the theoretical ?2 quantiles, which in the ideal
case gives a straight line. Such a quantile-quantile plot is constructed as follows: Let
?(i) denote the observed Mahalanobis distances ordered from lowest to highest, and p i the
cumulative proportion before each ?(i) given by pi = (i ? 1/2)/n. Let further zi = F ?1 pi
denote the theoretical quantile at position pi , where F is the cumulative ?2 -distribution
function. The quantile-quantile plot is then obtained by plotting ?(i) against zi . Deviations
from linearity can be formalized by fitting a linear model on the observed quantiles and
calculating confidence intervals around the fit. Observations falling outside the confidence
interval are then treated as outliers. A potential problem of this approach is that the outliers
themselves heavily influence the quantile-quantile fit. In order to overcome this problem,
the use of robust fitting procedures has been proposed in the literature, see e.g. [4]. In
the experiments below we use an M-estimator with Huber loss function. For estimating
confidence intervals around the fit we use the standard formula (see [2, 5])
p
?(?(i) ) = b ? (?2 (zi ))?1 (pi (1 ? pi ))/n,
(7)
which can be intuitively understood as the product of the slope b and the standard error of
the quantiles. A 100(1 ? ?)% envelope around the fit is then defined as ?(i) ? z?/2 ?(?(i) )
where z?/2 is the 1 ? (1 ? ?)/2 quantile of the standard normal distribution.
The choice of the confidence level ? is somewhat arbitrary, and from a conceptual viewpoint
one might even argue that the problem of specifying one free parameter (i.e. the expected
fraction of outliers) has simply been transferred into the problem of specifying another
one. In practice, however, selecting ? is a much more intuitive procedure than guessing the
fraction of outliers. Whereas the latter requires problem-specific prior knowledge which is
hardly available in practice, the former depends only on the variance of a linear model fit.
Thus, ? can be specified in a problem independent way.
4
Model selection
In our model the data are first mapped into some feature space, in which then a Gaussian
model is fitted. Mahalanobis distances to the mean of this Gaussian are computed by
evaluating (4). The feature space mapping is implicitly defined by the kernel function, for
which we assume that it is parametrized by a kernel parameter ?. For selecting all free
parameters in (4), we are, thus, left with the problem of selecting ? = (?, ?)> .
The idea is now to select ? by maximizing the cross-validated likelihood. From a theoretical viewpoint, the cross-validated (CV) likelihood framework is appealing, since in [13]
the CV likelihood selector has been shown to asymptotically perform as well as the optimal benchmark selector which characterizes the best possible model (in terms of KullbackLeibler divergence to the true distribution) contained in the parametric family.
For kernels that map into a space with dimension p > n, however, two problems arise: (i)
the subspace spanned by the mapped samples varies with different sample sizes; (ii) not
the whole feature space is accessible for vectors in the input space. As a consequence, it
is difficult to find a ?proper? normalization of the Gaussian density in the induced feature
space. We propose to avoid this problem by considering the likelihood in the input space
rather than in the feature space, i.e. we are looking for a properly normalized density model
p(x|?) in Rd such that p(x|?) has the same contour lines as the Gaussian model in the feature
space: p(xi |?) = p(xi |?) ? p(?(xi )|?) = p(?(xj )|?). Denoting by Xn = {xi }ni=1 a
sample from p(x) from which the kernel matrix K is built, a natural input space model is
R
pn (x|Xn , ?) = Z ?1 exp{? 21 D(x; Xn , ?)}, with Z = Rd pn (x|Xn , ?) dx, (8)
where D(x; Xn , ?) denotes the (parametrized) Mahalanobis distances (4) of a Gaussian
model in the feature space. Note that this density model in the input space has the same
form as our Gaussian model in the feature space, except for the different normalization
constant Z. Computing this constant Z requires us to solve a normalization integral over the
whole d-dimensional input space. Since in general this integral is not analytically tractable
for nonlinear kernel models, we propose to approximate Z by a Monte Carlo sampling
method. In our experiments, for instance, the VEGAS algorithm [6], which implements a
mixed importance-stratified sampling approach, showed to be a reasonable method for up
to 10 input dimensions.
By using the CV likelihood framework we are guaranteed to (asymptotically) perform as
well as the best model in the parametrized family. Thus, the question arises whether the
family of densities defined by a Gaussian model in a kernel-induced feature space is ?rich
enough? such that no systematic errors occur. For RBF kernels, the following lemma provides a positive answer to this question.
Lemma 2. Let k(xi , xj ) = exp(?kxi ? xj k2 /?). As ? ? 0 , pnP
(x|Xn , ?) converges to
n
a Parzen window with vanishing kernel width: pn (x|Xn , ?) ? n1 i=1 ?(x ? xi ).
A formal proof is omitted due to space limitations. The basic ingredients of the proof
are: (i) In the limit ? ? 0 the expansion coefficients approach ?
? ? 1/(1 + ?)1. Thus,
y? = K ?
? ? 1/(1 + ?)1 and s2 ? 1/(1 + ?). (ii) D(x; ?, ?) ?
PnC(x) < ?, if x ?
{xi }ni=1 , and D(x; ?, ?) ? ?, else. Finally pn (x|Xn , ?, ?) ? n1 i=1 ?(x ? xi ).
Note that in the limit ? ? 0 a Parzen window becomes an unbiased estimator for any
continuous density, which provides an asymptotic justification for our approach: the crossvalidated likelihood framework guarantees us to convergence to a model that performs as
well as the best model in our model class as n ? ?. The latter, however, is ?rich enough?
in the sense that it contains models which in the limit ? ? 0 converge to an unbiased
estimator for every continuous p(x). Since contour lines of pn (x) are contour lines of a
Gaussian model in the feature space, the Mahalanobis distances are expected to follow a
?2 distribution, and atypical objects can be detected by observing the distribution of the
empirical Mahalanobis distances as described in the last section.
It remains to show that describing the data as a Gaussian in a kernel-induced feature space
is a statistically sound model. This is actually the case, since there exist decay rates for the
kernel width ? such that n grows at a higher rate as the effective degrees of freedom df :
Lemma 3. Let k(xi , xj ) = exp(?kxi ? xj k2 /?) and pn (x|Xn , ?, ?) defined by (8). If
? ? 1 decays like O(n?1/2 ), and for fixed ? ? 1, the ratio df /n ? 0 as n ? ?.
A formal proof is omitted due to space limitations. The basic ingredients of the proof are:
? i as n ? ?, (ii) the eigenvalue spectrum of
(i) the eigenvalues ?0i of (1/n)K converge to ?
? i ? exp(??i2 ), (iii) for
a Gaussian RBF kernel decays atPan exponential-quadratic rate: ?
n
?1/2 2
?i )] ? n1/2 ? ?1 log(n/?)
n sufficiently large, it holds that i=1 1/[1 + (?/n) exp(n
(proof by induction, using the fact that ln(n + 1) ? ln(n) ? 1/(n2 + n) which follows
from a Taylor expansion of the logarithm) ? df (n)/n ? 0.
5
Experiments
The performance of the proposed method is demonstrated for an outlier detection task in the field of face recognition.
The Olivetti face database (see
http://www.uk.research.att.com/facedatabase.html) contains ten different images of each of
40 distinct subjects, taken under different lighting conditions and at different facial expressions and facial details (glasses / no glasses). None of the subjects, however, wears
sunglasses. All the images are taken against a homogeneous background with the subjects
in an upright, frontal position. In this experiment we additionally corrupted the dataset by
including two images in which we have artificially changed normal glasses to ?sunglasses?
as can be seen in figure 1. The goal is to demonstrate that the proposed method is able to
identify these two atypical images without any problem-dependent prior assumptions.
Figure 1: Original and corrupted images with in-painted ?sunglasses?.
Each of the 402 images is characterized by a 10-dimensional vector which contains the
projections onto the leading 10 eigenfaces (eigenfaces are simply the eigenvectors of the
images treated as pixel-wise vectorial objects). These vectors are feed into a RBF kernel
of the form k(xi , xj ) = exp(?kxi ? xj k2 /?). In a first step, the free model parameters
(?, ?) are selected by maximizing the cross-validated likelihood. A simple 2-fold cross
validation scheme is used: the dataset is randomly split into a training set and a test set of
equal size, the model is build from the training set (including the numerical solution of the
normalization integral), and finally the likelihood is evaluated on the test set. This procedure is repeated for different values of (?, ?). In order to simplify the selection process
we kept ? = 10?4 fixed and varied only ?. Both the test likelihood and the corresponding
model complexity measured in terms of the effective degrees of freedom (df ) are plotted
in figure 2. One can clearly identify both an overfitting and an underfitting regime, separated by a broad plateau of models with similarly high likelihood. The df -curve, however,
shows a similar plateau, indicating that all these models have comparable complexity. This
observation suggests that the results should be rather insensitive to variations of ? over values contained in this plateau. This suggestion is indeed confirmed by the results in figure
2, where we compared the quantile-quantile plot for the maximum likelihood parameter
value with that of a slightly suboptimal model. Both quantile plots look very similar, and
in both cases two objects clearly fall outside a 99% envelope around the linear fit. Outside the plateau (no figure due to space limitations) the number of objects considered as
outlies drastically increases in overfitting regime (? too small), or decreases to zero in the
underfitting regime (? too large).
In figure 3 again the quantile plot for the most likely model is depicted. This time, however,
both objects identified as outliers are related to the corresponding original images, which
in fact are the artificially corrupted ones. In addition, the uncorrupted images are localized
in the plot, indicating that they look rather typical.
Some implementation details. Presumably the easiest way of implementing the model is
to carry
Pout an eigenvalue decomposition of K. Both the the effective degrees of freedom
df = i ?i /(?i + ?) and the Mahalanobis distances in eq. (4) can then by derived easily
10
?400
20
?450
5
10
?500
30
? (i)
test log?likelihood
? (i)
30
?350
20
15
?300
10
40
?250
effective degrees of freedom
40
50
?200
?550
?600
5
10
15
20
25
? 2 quantiles
30
6
7
8
9
log(?)
10
11
12
13
5
10
15
20
25
?2 quantiles
30
50
Figure 2: Middle panel: Selecting the kernel width ? by cross-validated likelihood (solid
line). The dotted line shows the corresponding effective degrees of freedom (df ). Left +
right panels: quantile plot for optimal model (left) and slightly suboptimal model (right).
40
99.99%
20
30
? (i)
99%
99%
10
99.99%
5
10
15
? 2 quantiles
20
25
30
Figure 3: Quantile plot with linear fit (solid) and envelopes (99% and 99.99 %, dashed).
from this decomposition (see (5) and (6)). Efficient on-line variants can be implemented by
using standard update formulas for matrix inversion by partitioning. For an implementation
of the VEGAS algorithm see [9]. The R package ?car? provides a comfortable implementation of quantile-quantile plots and robust line fitting (see also http://www.R-project.org).
6
Conclusion
Detecting outliers by way of one-class classifiers aims at finding a boundary that separates
?typical? objects in a data sample from the ?atypical? ones. Standard approaches of this
kind suffer from the problem that they require prior knowledge about the expected fraction
of outliers. For the purpose of outlier detection, however, the availability of such prior
information seems to be an unrealistic (or even contradictory) assumption. The method
proposed in this paper overcomes this shortcoming by using a one-class KFD classifier
which is directly related to Gaussian density estimation in the induced feature space. The
model benefits from both the built-in classification method and the explicit parametric density model: from the former it inherits the simple complexity regulation mechanism based
on only two tuning parameters. Moreover, within the classification framework it is possible to quantify the model complexity in terms of the effective degrees of freedom df . The
Gaussian density model, on the other hand, makes it possible to derive a formal description of atypical objects by way of hypothesis testing: Mahalanobis distances are expected
to follow a ?2 -distribution in df dimensions, and deviations from this distribution can be
quantified by confidence intervals around a fitted line in a quantile-quantile plot. Since
the density model is parametrized by both the kernel function and the regularization constant, it is necessary to select these free parameters before the outlier detection phase. This
parameter selection is achieved by observing the cross-validated likelihood for different
parameter values, and choosing those parameters which maximize this quantity. The theoretical motivation for this selection process follows from [13] where it has been shown
that the cross-validation selector asymptotically performs as well as the so called benchmark selector which selects the best model contained in the parametrized family of models.
Moreover, for RBF kernels it is shown in lemma 2 that the corresponding model family is
?rich enough? in the sense that it contains an unbiased estimator for the true density (as
long as it is continuous) in the limit of vanishing kernel width. Lemma 3 shows that there
exist decay rates for the kernel width such that the ratio of effective degrees of freedom and
sample size approaches zero.
The experiment on detecting persons wearing sunglasses within a collection of rather heterogeneous face images effectively demonstrates that the proposed method is able to detect
atypical objects without prior assumptions on the expected number of outliers. In particular, it demonstrates that the whole processing pipeline consisting of model selection by
cross-validated likelihood, fitting linear quantile-quantile models and detecting outliers by
considering confidence intervals around the fit works very well in practical applications
with reasonably small input dimensions. For input dimensions 10 the numerical solution of the normalization integral becomes rather time consuming when using the VEGAS
algorithm. Evaluating the usefulness of more sophisticated sampling models like MarkovChain Monte-Carlo methods for this particular task will be subject of future work.
Acknowledgments. The author would like to thank Tilman Lange, Mikio Braun and
Joachim M. Buhmann for helpful discussions and suggestions.
References
[1] R. Duda, P. Hart, and D. Stork. Pattern Classification. Wiley & Sons, 2001.
[2] J. Fox. Applied Regression, Linear Models, and Related Methods. Sage, 1997.
[3] T. Hastie, A. Buja, and R. Tibshirani. Penalized discriminant analysis. Annals of Statistics,
23:73?102, 1995.
[4] P.J. Huber. Robust Statistics. Wiley, 1981.
[5] M. Kendall and A. Stuart. The Advanced Theory of Statistics, volume 1. McMillan, 1977.
[6] G.P. Lepage. Vegas: An adaptive multidimensional integration program. Technical Report
CLNS-80/447, Cornell University, 1980.
[7] S. Mika, G. R?atsch, J. Weston, B. Sch?olkopf, and K.-R. M?uller. Fisher discriminant analysis
with kernels. In Y.-H. Hu, J. Larsen, E. Wilson, and S. Douglas, editors, Neural Networks for
Signal Processing IX, pages 41?48. IEEE, 1999.
[8] J. Moody. The effective number of parameters: An analysis of generalisation and regularisation
in nonlinear learning systems. In J. Moody, S. Hanson, and R. Lippmann, editors, NIPS 4, 1992.
[9] W.H. Press, S.A. Teukolsky, W.T. Vetterling, and B.P. Flannery. Numerical Recipies in C.
Cambridge University Press, 1992.
[10] V. Roth and V. Steinhage. Nonlinear discriminant analysis using kernel functions. In S.A. Solla,
T.K. Leen, and K.-R. M?uller, editors, NIPS 12, pages 568?574. MIT Press, 2000.
[11] B. Sch?olkopf, S. Mika, C. Burges, P. Knirsch, K.-R. Mu? ller, G. R?atsch, and A. Smola. Input
space vs. feature space in kernel-based methods. IEEE Trans. Neural Networks, 10(5), 1999.
[12] B. Sch?olkopf, R.C. Williamson, A. Smola, and J. Shawe-Taylor. SV estimation of a distribution?s support. In S. Solla, T. Leen, and K.-R. Mu? ller, editors, NIPS 12, pages 582?588. 2000.
[13] M.J. van der Laan, S. Dudoit, and S. Keles. Asymptotic optimality of likelihood-based crossvalidation. Statistical Applications in Genetics and Molecular Biology, 3(1), 2004.
| 2656 |@word middle:1 inversion:1 seems:3 proportion:1 nd:3 duda:1 hu:1 covariance:2 decomposition:2 solid:2 carry:1 contains:6 att:1 selecting:5 denoting:4 com:1 scatter:2 dx:1 must:1 numerical:3 plot:12 update:1 v:1 selected:2 vanishing:2 detecting:7 provides:4 characterization:2 org:1 constructed:1 fitting:5 underfitting:2 pnp:1 huber:2 indeed:1 expected:9 themselves:1 chi:1 window:2 considering:2 becomes:2 project:1 estimating:6 underlying:1 moreover:3 maximizes:1 linearity:1 panel:2 lowest:1 easiest:1 kind:1 eigenvector:1 textbook:1 finding:2 transformation:2 guarantee:1 every:1 multidimensional:1 braun:1 classifier:9 k2:4 scaled:1 control:2 normally:1 uk:1 partitioning:1 demonstrates:2 comfortable:1 positive:2 before:2 understood:1 limit:4 consequence:1 painted:1 might:2 plus:1 mika:2 quantified:1 specifying:3 suggests:1 stratified:1 statistically:1 practical:2 acknowledgment:1 testing:1 practice:3 implement:1 procedure:4 empirical:1 eth:1 reject:1 projection:2 confidence:6 onto:2 close:2 selection:7 impossible:1 influence:1 restriction:1 equivalent:1 map:1 demonstrated:1 roth:2 missing:1 maximizing:2 www:2 formalized:1 estimator:6 spanned:1 notion:1 coordinate:1 justification:1 variation:1 annals:1 heavily:2 homogeneous:1 hypothesis:3 origin:1 recognition:1 labeled:2 database:1 observed:4 calculate:1 solla:2 decrease:2 highest:1 mu:2 complexity:7 depend:1 solving:1 negatively:1 completely:1 easily:2 represented:1 separated:1 distinct:1 shortcoming:2 effective:10 monte:2 detected:1 outside:3 choosing:1 solve:1 relax:1 statistic:4 jointly:1 advantage:1 eigenvalue:5 propose:3 product:1 realization:1 rapidly:1 intuitive:2 description:1 ky:1 olkopf:3 crossvalidation:1 convergence:1 converges:1 object:19 help:1 derive:4 measured:1 pnc:1 eq:2 implemented:1 come:3 quantify:1 differ:1 direction:1 closely:1 centered:1 implementing:1 require:1 hold:1 sufficiently:3 considered:4 around:6 normal:2 exp:6 presumably:1 mapping:1 omitted:2 purpose:1 estimation:10 establishes:1 weighted:4 uller:2 mit:1 clearly:2 gaussian:25 aim:2 rather:6 avoid:2 pn:6 cornell:1 volker:1 wilson:1 validated:7 inherits:2 derived:1 joachim:1 properly:1 likelihood:18 detect:3 sense:4 glass:3 helpful:1 dependent:1 membership:1 i0:2 facedatabase:1 unlikely:1 typically:1 vetterling:1 kernelized:1 relation:4 transformed:1 selects:1 pixel:1 arg:1 classification:9 html:1 priori:1 constrained:1 special:1 integration:1 field:1 equal:1 having:1 sampling:3 biology:1 broad:1 look:4 unsupervised:1 stuart:1 future:1 minimized:1 report:1 simplify:1 inherent:2 randomly:1 divergence:1 phase:1 consisting:1 n1:3 attempt:1 freedom:11 detection:6 interest:1 kfd:5 highly:1 severe:1 integral:4 necessary:2 facial:2 orthogonal:4 fox:1 euclidean:2 taylor:2 logarithm:1 desired:1 plotted:1 theoretical:5 fitted:3 instance:1 introducing:1 deviation:5 pout:1 usefulness:1 too:3 kullbackleibler:1 answer:1 varies:1 corrupted:3 sv:1 kxi:3 person:1 density:23 fundamental:1 accessible:1 systematic:1 parzen:2 moody:2 squared:1 again:2 central:1 opposed:1 knirsch:1 leading:1 account:1 potential:1 pooled:1 availability:1 coefficient:2 depends:1 view:1 kendall:1 observing:3 characterizes:1 start:1 complicated:1 slope:1 contribution:3 square:2 ni:3 variance:1 identify:3 none:1 carlo:2 lighting:1 confirmed:1 straight:1 plateau:4 against:3 larsen:1 proof:5 dataset:3 knowledge:2 car:1 formalize:2 sophisticated:1 actually:1 feed:1 higher:1 follow:2 leen:2 evaluated:1 generality:1 smola:2 hand:2 nonlinear:4 defines:1 lda:7 grows:1 concept:1 unbiased:4 true:5 counterpart:1 former:2 regularization:3 analytically:2 normalized:1 read:2 i2:1 mahalanobis:13 width:5 oc:4 coincides:1 criterion:1 ridge:4 demonstrate:1 performs:2 reasoning:1 image:10 wise:1 vega:4 recently:1 discriminants:1 functional:3 stork:1 insensitive:1 volume:1 linking:1 interpretation:1 relating:1 cambridge:1 cv:3 rd:4 tuning:1 similarly:1 replicating:1 shawe:1 wear:1 showed:1 olivetti:1 inf:1 scenario:1 certain:2 der:1 uncorrupted:1 seen:1 somewhat:1 converge:2 maximize:1 ller:2 dashed:1 ii:3 signal:1 sound:1 reduces:1 technical:2 characterized:1 cross:10 long:1 hart:1 molecular:1 variant:1 regression:5 basic:2 heterogeneous:1 essentially:2 df:15 kernel:34 normalization:6 achieved:1 justified:2 whereas:1 background:1 addition:1 interval:5 else:1 sch:3 envelope:3 rest:1 strict:1 probably:1 induced:7 subject:4 steinhage:1 ideal:1 intermediate:1 iii:1 enough:4 split:1 xj:8 fit:8 zi:3 hastie:1 identified:2 opposite:1 suboptimal:2 lange:1 idea:3 translates:1 whether:1 expression:1 penalty:1 suffer:1 hardly:2 useful:1 clear:1 eigenvectors:3 amount:1 ten:1 simplest:1 http:2 exist:2 notice:1 shifted:1 dotted:1 sign:1 correctly:1 tibshirani:1 falling:1 douglas:1 kept:1 asymptotically:4 fraction:6 sum:1 package:1 arrive:1 family:5 reasonable:1 decision:1 scaling:2 comparable:1 bound:1 guaranteed:1 fold:1 quadratic:1 occur:1 vectorial:1 constraint:1 keles:1 min:1 optimality:1 transferred:1 according:1 slightly:3 son:1 appealing:2 making:1 outlier:32 intuitively:1 taken:2 pipeline:1 ln:2 equation:1 zurich:2 remains:1 describing:1 mechanism:4 needed:1 tractable:1 available:3 observe:1 away:1 appropriate:2 rp:1 original:3 assumes:1 denotes:2 graphical:1 calculating:1 restrictive:1 quantile:23 build:1 classical:1 objective:1 question:2 quantity:4 parametric:2 usual:1 guessing:1 subspace:4 distance:19 thank:1 unable:1 separate:3 separating:4 majority:1 mapped:3 link:1 parametrized:5 topic:1 argue:1 considers:1 discriminant:6 induction:1 assuming:1 ratio:2 minimizing:1 regulation:2 difficult:1 trace:2 negative:2 sage:1 implementation:3 proper:1 unknown:1 perform:2 observation:2 benchmark:2 clns:1 tilman:1 defining:1 looking:1 rn:1 varied:1 arbitrary:1 mcmillan:1 buja:1 overcoming:1 introduced:1 specified:3 hanson:1 established:1 nip:3 trans:1 address:1 able:2 usually:2 below:1 pattern:1 regime:3 program:1 built:2 including:2 unpenalized:2 unrealistic:1 suitable:2 natural:2 treated:2 indicator:1 buhmann:1 advanced:1 scheme:1 sunglass:4 carried:1 prior:8 literature:1 asymptotic:2 regularisation:1 loss:2 squeezed:1 mixed:1 interesting:1 limitation:3 suggestion:2 ingredient:3 localized:1 validation:3 degree:11 mercer:1 plotting:1 viewpoint:2 editor:4 pi:5 row:1 genetics:1 penalized:3 changed:1 last:1 free:6 drastically:1 formal:4 side:1 burges:1 institute:1 fall:1 eigenfaces:2 face:3 crossvalidated:1 distributed:2 benefit:2 overcome:3 boundary:7 dimension:7 xn:10 cumulative:2 rich:4 contour:5 evaluating:2 curve:1 collection:1 author:1 replicated:1 projected:2 adaptive:1 far:1 approximate:1 selector:4 lippmann:1 implicitly:2 overcomes:2 overfitting:2 conceptual:2 conclude:1 consuming:1 xi:15 spectrum:1 continuous:3 additionally:2 reasonably:1 robust:4 expansion:3 williamson:1 artificially:2 diag:2 main:4 whole:4 s2:5 arise:1 motivation:1 n2:1 repeated:1 x1:1 augmented:1 mikio:1 quantiles:7 wiley:2 inferring:1 position:2 explicit:3 exponential:1 lie:2 atypical:12 ix:1 down:2 formula:2 specific:1 decay:4 svm:3 derives:1 consist:1 adding:1 effectively:2 importance:2 flannery:1 depicted:1 simply:3 likely:1 unexpected:2 ordered:2 contained:3 partially:1 van:1 ch:2 determines:1 constantly:1 teukolsky:1 weston:1 goal:1 dudoit:1 quantifying:1 rbf:6 fisher:4 hard:1 typical:3 except:1 upright:1 generalisation:1 laan:1 lemma:9 contradictory:1 total:1 called:1 atsch:2 indicating:2 select:3 support:1 latter:4 fulfills:1 arises:1 ethz:1 frontal:1 wearing:1 |
1,824 | 2,657 | Class-size Independent Generalization Analsysis
of Some Discriminative Multi-Category
Classification Methods
Tong Zhang
IBM T.J. Watson Research Center
Yorktown Heights, NY 10598
[email protected]
Abstract
We consider the problem of deriving class-size independent generalization bounds for some regularized discriminative multi-category classification methods. In particular, we obtain an expected generalization
bound for a standard formulation of multi-category support vector machines. Based on the theoretical result, we argue that the formulation over-penalizes misclassification error, which in theory may lead to
poor generalization performance. A remedy, based on a generalization
of multi-category logistic regression (conditional maximum entropy), is
then proposed, and its theoretical properties are examined.
1
Introduction
We consider the multi-category classification problem, where we want to find a predictor
p : X ? Y, where X is a set of possible inputs and Y is a discrete set of possible outputs.
In many applications, the output space Y can be extremely large, and may be regarded as
infinity for practical purposes. For example, in natural language processing and sequence
analysis, the input can be an English sentence, and the output can be a parse or a translation
of the sentence. For such applications, the number of potential outputs can be exponential
of the length of the input sentence. As another example, in machine learning based webpage search and ranking, the input is the keywords and the output space consists of all
web-pages.
In order to handle such application tasks, from the theoretical point of view, we do not
need to assume that the output space Y is finite, so that it is crucial to obtain generalization
bounds that are independent of the size of Y. For such large scale applications, one often
has a routine that maps each x ? X to a subset of candidates GEN(x) ? Y, so that the
desired output associated with x belongs to GEN(x). For example, for web-page search,
GEN(x) consists of all pages that contain one or more keywords in x. For sequence
annotation, GEN(x) may include all annotation sequences that are consistent. Although
the set GEN(x) may significantly reduce the size of potential outputs Y, it can still be
large. Therefore it is important that our learning bounds are independent of the size of
GEN(x).
We consider the general setting of learning in Hilbert spaces since it includes the popular
kernel methods. Let our feature space H be a reproducing kernel Hilbert space with dot
product ?. For a weight vector w ? H, we use notation kwk2H = w ? w. We associate each
possible input/output pair (x, y) ? X ? Y with a feature vector fx,y ? H. Our classifier is
characterized by a weight vector w ? H, with the following classification rule:
pw (x) = arg
max
c?GEN(x)
w ? fx,c .
(1)
Note that computational issues are ignored in this paper. In particular, we assume that the
above decision can be computed efficiently (either approximately or exactly) even when
GEN(x) is large. In practice, this is often possible either by heuristic search or dynamic
programming (when GEN(X) has certain local-dependency structures). In this paper, we
are only interested in the learning performance, so that we will not discuss the computational aspect.
We assume that the input/output pair (x, y) ? X ?Y is drawn from an unknown underlying
distribution D. The quality of the predictor w is measured by some loss function. In this
paper, we focus on the expected classification error with respect to D:
`D (w) = E(X,Y ) I(pw (X), Y ),
(2)
where (X, Y ) is drawn from D, and I is the standard 0-1 classification error: I(Y 0 , Y ) = 0
when Y 0 = Y and I(Y 0 , Y ) = 1 when Y 0 6= Y .
The general set up we described above is useful for many application problems, and has
been investigated, for example, in [2, 6]. The important issue of class-size independent (or
weakly dependent) generalization analysis has also been discussed there.
Consider a set of training data S = {(Xi , Yi ), i = 1, . . . , n}, where we assume that for
? S ? H such that the classification error
each i, Yi ? GEN(Xi ). We would like to find w
? S ) is as small as possible. This paper studies regularized discriminative learning
`D (w
? S ? H by solving the following optimization
methods that estimate a weight vector w
problem:
#
" n
?
1X
2
? S = arg min
L(w, Xi , Yi ) + kwkH ,
(3)
w
w?H n
2
i=1
where ? ? 0 is an appropriately chosen regularization parameter, and L(w, X, Y ) is a
loss function which is convex of w. In this paper, we focus on some loss functions of the
following form:
?
?
X
L(w, X, Y ) = ? ?
?(w ? (fX,Y ? fX,c ))? ,
c?GEN(X)\Y
where ? and ? are appropriately chosen real-valued functions.
Typically ? is chosen as an increasing function and ? as a decreasing function, selected so that (3) is a convex optimization problem. The intuition behind this method
is that the resulting optimization formulation favors large values w ? (fXi ,Yi ? fXi ,c )
for all c ? GEN(Xi )\Yi . Therefore, it favors a weight vector w ? H such that
w ? fXi ,Yi = arg maxc?GEN(Xi ) w ? fXi ,c , which encourages the correct classification
rule in (1). The regularization term ?2 kwk2H is included for capacity control, which has
become the standard practice in machine learning nowadays.
Two of the most important methods used in practice, multi-category support vector machines [7] and penalized multi-category logistic regression (conditional maximum entropy
with Gaussian smoothing [1]), can be regarded as special cases of (3). The purpose of this
paper is to study their generalization behaviors. In particular, we are interested in generalization bounds that are independent of the size of GEN(Xi ).
2
Multi-category Support Vector Machines
We consider the multi-category support vector machine method proposed in [7]. It is a
? S computed based on the following formula:
special case of (3) with w
?
?
n
X
X
?
1
? S = arg min ?
h(w ? (fXi ,Yi ? ?fXi ,c )) + kwk2H ? ,
(4)
w
w?H n
2
i=1
c?GEN(Xi )\Yi
where h(z) = max(1 ? z, 0) is the hinge loss used in the standard SVM formulation.
From the asymptotic statistical point of view, this formulation has some drawbacks in that
there are cases such that the method does not lead to a classifier that achieves the Bayes
error [9] (inconsistency). A Bayes consistent remedy has been proposed in [4]. However,
method based on (4) has some attractive properties, and has been successfully used for
some practical problems.
We are interested in the generalization performance of (4). As we shall see, this formulation
performs very well in the linearly separable (or near separable) case. Our analysis also
reveals a problem of this method for non-separable problems. Specifically, the formulation
over-penalizes classification error. Possible remedies will be suggested at the end of the
section.
We start with the following theorem, which specifies a generalization bound in a form often
referred to as the oracle inequality. That is, it bounds the generalization performance of
the SVM method (4) in terms of the best possible true multi-category SVM loss. Proof is
left to Appendix B.
Theorem 2.1 Let M = supX supY,Y 0 ?GEN(X) kfX,Y ? fX,Y 0 kH . The expected generalization error of (4) can be bounded as:
? S ) ? ES E(X,Y )
ES `D (w
? S ? (fX,Y ? fX,c ))
h(w
sup
c?GEN(X)\Y
2
?
max(?n, M ) + M
?n
?
2
inf ?E(X,Y )
w?H
X
h(w ? (fX,Y ? fX,c )) +
c?GEN(X)\Y
?nkwk2H
2(n + 1)
?
?,
where ES is the expectation with respect to the training data.
Note that the generalization bound does not depend on the size of GEN(X), which is
what we want to achieve. The left-hand side of the theorem bounds the classification error
? S ? (fX,Y ? fX,c )),
of the multi-category SVM classifier in terms of supc?GEN(X)\Y h(w
P
while the right hand side in terms of c?GEN(X)\Y h(w ? (fX,Y ? fX,c )). There is a
mismatch here. The latter is a very loose bound since it over-counts classification errors
in the summation when multiple errors are made at the same point. In fact, although the
class-size dependency does
Pnot come into our generalization analysis, it may well come
into the summation term c?GEN(X)\Y h(w ? (fX,Y ? fX,c )) when multiple errors are
made at the same point. We believe that this is a serious flaw of the method, which we
will try
Pto remedy later. However, the bound can be quite tight in the near separable case,
? S ? (fX,Y ? fX,c )) is small. The following Corollary gives such
when c?GEN(X)\Y h(w
a result:
Corollary 2.1 Assume that there is a large margin separator w? ? H such that for each
data point (X, Y ), the following margin condition holds:
?c ? GEN(X)\Y : w? ? fX,Y ? w? ? fX,c + 1.
Then in the limit of ? ? 0, the expected generalization error of (4) can be bounded as:
? S) ?
ES `D (w
kw? k2H
sup
sup
kfX,Y ? fX,Y 0 k2H ,
n + 1 X Y,Y 0 ?GEN(X)
where ES is the expectation with respect to the training data.
Proof. Just choose w? on the right hand side of Theorem 2.1. 2
The above result for (4) gives a class-size independent bound for large margin separable
problems. The bound generalizes a similar result for two-class hard-margin SVM. It also
matches a bound for multi-class perceptron in [2]. To our knowledge, this is the first result
showing that the generalization performance of a batch large margin algorithm such as (4)
can be class-size independent (at least in the separable case). Previous results in [2, 6],
relying on the covering number analysis, lead to bounds that depend on the size of Y
(although the result in [6] is of a different style).
Our analysis also implies that the multi-category classification method (4) has good generalization behavior for separable problems. However, as pointed out earlier, for nonseparable problems, the formulation over-penalize classification error since in the summation, it may count classification error at a point multiple times when multiple
mistakes
P
are made at the point. A remedy is to replace the summation symbol c?GEN(Xi )\Yi
in (4) by the sup operator supc?GEN(Xi )\Yi , as we have used for bounding the classification error on the left hand side of Theorem 2.1. This is done in [3]. However, like
(4), the resulting formulation is also inconsistent. Instead of using a hard-sup operator,
we may also use a soft-sup operator, which can possibly
lead to consistency. For examP
ple, consider the equality supc |hc | = limp?? ( c |hc |p )1/p , we may approximate the
right hand side limit with a P
large p. Another more interesting formulation is to consider
supc hc = limp?? p?1 ln( c exp(phc )), which leads to a generalization of the conditional maximum entropy method.
3
Large Margin Discriminative Maximum Entropy Method
Based on the motivation given at the end of the last section, we propose the following generalization of maximum entropy (multi-category logistic regression) with Gaussian prior
(see [1]). It introduces a margin parameter into the standard maximum entropy formulation, and can be regarded as a special case of (3):
?
n
X
1
1
? S = arg min ?
ln ?1 +
w
w?H
n i=1 p
?
?
X
c?GEN(Xi )\Yi
p(??w?(fXi ,Yi ?fXi ,c )) ?
e
?
?
2 ?
+ kwkH ,
2
(5)
where ? is a margin condition, and p > 0 is a scaling factor (which in theory can also be
removed by a redefinition of w and ?).
If we choose ? = 0, then this formulation is equivalent to the standard maximum entropy
method. If we pick the margin parameter ? = 1, and let p ? ?, then
?
?
X
1 ?
ln 1 +
ep(??w?(fXi ,Yi ?fXi ,c )) ? ?
sup
h(w?(fXi ,Yi ?fXi ,c )),
p
c?GEN(Xi )\Yi
c?GEN(Xi )\Yi
where h(z)
P = max(0, 1 ? z) is used in (4). In this case, the formulation reduces to (4)
but with c?GEN(Xi )\Yi replaced by supc?GEN(Xi )\Yi . As discussed at the end of last
section, this solves the problem of over-counting the classification error.
In general, even with a finite scaling factor p, the log-transform in (4) guarantees that
one penalizes misclassification error at most p1 ln |GEN(Xi )| times at a point, where
|GEN(Xi )| is the size of GEN(Xi ), while in (4), one may potentially over-penalize
|GEN(Xi )| times. Clearly this is a desirable effect for non-separable problems. Methods in (5) have many attractive properties. In particular, we are able to derive class-size
independent generalization bounds for this method. The proof of the following theorem is
given in Appendix C.
Theorem 3.1 Let M = supX supY,Y 0 ?GEN(X) kfX,Y ? fX,Y 0 kH . Define loss L(w, x, y)
as:
?
?
X
1
ep(??w?(fx,y ?fx,c )) ? ,
L(w, x, y) = ln ?1 +
p
c?GEN(x)\y
and let
E(X,Y ) L(w, X, Y ) +
?n
kwk2H .
w?H
2(n + 1)
The expected generalization error of (5) can be bounded as:
Q? = inf
? S , X, Y ) ? Q? +
ES E(X,Y ) L(w
M2
(1 ? e?pQ? ).
?n
where ES is the expectation with respect to the training data.
Theorem 3.1 gives a class-size independent generalization bound for (5). Note that the left
? S from (5), and the right hand size is specified in terms
hand side is the true loss of the w
of the best possible regularized true loss Q? , plus a penalty term that is no larger than
M 2 /(?n). It is clear that this generalization bound is class-size independent. Moreover,
unlike Theorem 2.1, the loss function on the left hand side matches the loss function on
the right hand side in Theorem 3.1. These are not trivial properties. In fact, most learning
methods do not have these desirable properties. We believe this is a great advantage for the
maximum entropy-type discriminative learning method in (5). It implies that this class of
algorithms are suitable for problems with large number of classes. Moreover, we can see
that the generalization performance is well-behaved no matter what values of p and ? we
choose.
If we take ? = 0 and p = 1, then we obtain a generalization bound for the popular maximum entropy method with Gaussian prior, which has been widely used in natural language
processing applications. To our knowledge, this is the first generalization bound derived for
this method. Our result not only shows the importance of Gaussian prior regularization, but
also implies that the regularized conditional maximum entropy method has very desirable
generalization behavior.
Another interesting special case of (5) is to let ? = 1 and p ? ?. For simplicity we only
consider the case that |GEN(X)| is finite (but can be arbitrarily large). In this case, we
note that 0 ? L(w, X, Y ) ? supc?GEN(X)\Y h(w ? (fX,Y ? fX,c )) ? ln |GEN(X)|
. We
p
thus obtain from Theorem 3.1 a bound
ES E(X,Y )
c?GEN(X)\Y
"
+ inf
w?H
E(X,Y )
EX ln |GEN(X)| M 2
+
p
?n
#
?kwk2H
? fX,c )) +
.
2
? S ? (fX,Y ? fX,c )) ?
h(w
sup
sup
c?GEN(X)\Y
h(w ? (fX,Y
Now we can take a sufficiently large p such that the term EX ln |GEN(X)|/p becomes
negligible. Let p ? ?, the result implies a bound for the SVM method in [3]. For
non-separable problems, this bound is clearly superior
Pto the SVM bound in Theorem 2.1
since the right hand side replaces the summation c?GEN(X)\Y by the sup operator
supc?GEN(X)\Y . In theory, this satisfactorily solves the problem of over-penalizing misclassification error. Moreover, an advantage over [3] is that for some p, consistency can be
achieved. Our analysis also establishes a bridge between the Gaussian smoothed maximum
entropy method [1] and the SVM method in [3].
4
Conclusion
We studied the generalization performance of some regularized multi-category classification methods. In particular, we derived a class-size independent generalization bound for
a standard formulation of multi-category support vector machines. Based on the theoretical investigation, we showed that this method works well for linearly separable problems.
However, it over-penalizes mis-classification error, leading to loose generalization bounds
in the non-separable case. A remedy, based on a generalization of the maximum entropy
method, is proposed. Moreover, we are able to derive class-size independent bounds for the
newly proposed formulation, which implies that this class of methods (including the standard maximum entropy) are suitable for classification problems with very large number of
classes. We showed that in theory, the new formulation provides a satisfactory solution to
the problem of over-penalizing mis-classification error.
A
A general stability bound
The following lemma is essentially a variant of similar stability results for regularized learning systems used in [8, 10]. We include the proof Sketch for completeness.
Lemma A.1 Consider a sequence of convex functions Li (w) for i = 1, 2, . . . Define for
k = 1, 2, . . .
#
" k
X
?n
2
kwkH .
Li (w) +
wk = arg min
w
2
i=1
Then for all k ? 1, there exists subgradient (cf. [5]) ?Lk+1 (wk+1 ) of Li at wk+1 such
that
k+1
1 X
1
wk+1 = ?
?Li (wk+1 ), kwk ? wk+1 kH ?
k?Lk+1 (wk+1 )kH .
?n i=1
?n
Proof Sketch. The first equality is the first-order condition for the optimization problem [5]
where wk+1 is the solution. Now, subtracting this equality at wk and wk+1 , we have:
??n(wk+1 ? wk ) = ?Lk+1 (wk+1 ) +
k
X
(?Li (wk+1 ) ? ?Li (wk )).
i=1
Multiply the two sides by wk+1 ? wk , we obtain
??nkwk+1 ?wk k2H = ?Lk+1 (wk+1 )?(wk+1 ?wk )+
k
X
(?Li (wk+1 )??Li (wk ))?(wk+1 ?wk ).
i=1
Note that ?Li (wk+1 ) ? ?Li (wk )) ? (wk+1 ? wk ) = dLi (wk , wk+1 ) + dLi (wk+1 , wk ),
where dL (w, w0 ) = L(w0 ) ? L(w) ? ?L(w) ? (w0 ? w) is often called the Bregman
divergence of L, which is well-known to be non-negative for any convex function L (this
claim is also easy to verify by definition). We thus have (?Li (wk+1 )??Li (wk ))?(wk+1 ?
wk ) ? 0. It follows that
??nkwk+1 ? wk k2H ? ?Lk+1 (wk+1 ) ? (wk+1 ? wk ) ? ?k?Lk+1 (wk+1 )kH kwk+1 ? wk kH .
By canceling the factor kwk+1 ? wk kH , we obtain the second inequality. 2
B
Proof Sketch of Theorem 2.1
? k be the solution of (4) with
Consider training samples (Xi , Yi ) for i = 1, . . . , n + 1. Let w
Pn+1
the training sample (Xk , Yk ) removed from the set (that is, the summation is i=1,i6=k ),
P
P
? be the solution of (4) but with the summation ni=1 replaced by n+1
and let w
i=1 . Now for
? ? (fXk ,Yk ? fXk ,c ) for c ? GEN(X). It follows from
notation simplicity, we let zk,c = w
Lemma A.1 that
? 2H = ?
kwk
n+1
1 X
?n
X
h0 (zk,c )zk,c ,
? k ? wk
? H ??
kw
k=1 c?GEN(X)
M
?n
X
h0 (zk,c ),
c?GEN(X)\Y
0
where h (?) denotes a subgradient of h(?). Therefore using the inequality ?h0 (z) ? h(z) ?
h0 (z)z, we have
? k ? (fXk ,Yk ? fXk ,c )) ? h(zk,c )] ? kw
? k ? wk
? HM
[h(w
sup
c?GEN(Xk )\Yk
??
M2
?n
X
h0 (zk,c ) ?
c?GEN(Xk )\Yk
M2
?n
X
[h(zk,c ) ? h0 (zk,c )zk,c ].
c?GEN(Xk )\Yk
Summing over k = 1, . . . , n + 1, we obtain
n+1
X
? k ? (fXk ,Yk ? fXk ,c )) ? h(zk,c )]
[h(w
sup
k=1 c?GEN(Xk )\Yk
M2
?
?n
X
M
?n
[h(zk,c ) ? h0 (zk,c )zk,c ]
c?GEN(Xk )\Yk k=1
2
=
n+1
X
X
n+1
X
? 2H M 2 .
h(zk,c ) + kwk
c?GEN(Xk )\Yk k=1
Therefore given an arbitrary w ? H, we have
n+1
X
? k ? (fXk ,Yk ? fXk ,c ))
h(w
sup
k=1 c?GEN(Xk )\Yk
?(1 +
M2
)
?n
X
n+1
X
? 2H M 2
h(zk,c ) + kwk
c?GEN(Xk )\Yk k=1
?
?
n+1
X
X
M 2 2M 2 ?
?n
? 2H ?
)
h(zk,c ) +
kwk
? max(1 +
,
?n ?n
2
c?GEN(Xk )\Yk k=1
?
?
n+1
X
X
M 2 2M 2 ?
?n
? max(1 +
,
)
h(w ? (fXk ,Yk ? fXk ,c )) +
kwk2H ? .
?n ?n
2
c?GEN(Xk )\Yk k=1
Now, taking expectation with respect to the training data, we obtain the bound.
C
Proof Sketch of Theorem 3.1
Similar to the proof of Theorem 2.1, we consider training samples (Xi , Yi ) for i =
? k be the solution of (5) with the training sample (Xk , Yk ) removed
1, . . . , n + 1. Let w
Pn+1
? be the solution of (5) but with
from the set (that is, the summation is i=1,i6=k ), and let w
Pn
Pn+1
the summation i=1 replaced by i=1 . It follows from Lemma A.1 that
1
M
? k ,Yk )
? k ? wk
? H?
? Xk , Yk )kH ?
kw
k?L(w,
(1 ? e?pL(w,X
).
?n
?n
Therefore
M2
? k ,Yk )
? Xk , Yk ) ?
? k , Xk , Yk ) ? L(w,
(1 ? e?pL(w,X
).
L(w
?n
Now summing over k, we obtain
n+1
n+1
1 X
1 X
M2
? k , Xk , Yk ) ?
? Xk , Yk ) +
L(w
L(w,
n+1
n+1
?n
k=1
k=1
n+1
1 X ?pL(w,X
?
k ,Yk )
1?
e
n+1
!
.
k=1
Taking expectation with respect to the training data, and using the following Jensen?s
inequality:
n+1
Pn+1
1
1 X ?pL(w,X
? k ,Yk )
? k ,Yk )
,
?ES
e
? ?e?pES n+1 k=1 L(w,X
n+1
k=1
we obtain
k
? , Xk , Yk ) ? ES
ES E(Xk ,Yk ) L(w
n+1
X
k=1
Pn+1
? Xk , Yk ) M 2
L(w,
1 ? e?pES k=1
+
n+1
?n
?
L(w,X
k ,Yk )
n+1
Pn+1
? Xk , Yk ) ? (n+1)Q? (which follows from the optimal
Now, using the fact ES k=1 L(w,
? we obtain the theorem.
property of w),
References
[1] Stanley Chen and Ronald Rosenfeld. A survey of smoothing techniques for ME models. IEEE Trans. Speech and Audio Processing, 8:37?50, 2000.
[2] Michael Collins.
Parameter estimation for statistical parsing models: Theory and practice of distribution-free methods. In IWPT, 2001. available at
http://www.ai.mit.edu/people/mcollins/publications.html.
[3] Koby Crammer and Yoram Singer. On the algorithmic implementation of multiclass
kernel-based vector machines. Journal of Machine Learning Research, 2:265?292,
2001.
[4] Y. Lee, Y. Lin, and G. Wahba. Multicategory support vector machines, theory, and
application to the classification of microarray data and satellite radiance data. Journal
of American Statistical Association, 99:67?81, 2004.
[5] R. Tyrrell Rockafellar. Convex analysis. Princeton University Press, Princeton, NJ,
1970.
[6] Ben Taskar, Carlos Guestrin, and Daphne Koller. Max-margin markov networks.
In Sebastian Thrun, Lawrence Saul, and Bernhard Sch?olkopf, editors, Advances in
Neural Information Processing Systems 16. MIT Press, Cambridge, MA, 2004.
[7] J. Weston and C. Watkins. Multi-class support vector machines. Technical Report
CSD-TR-98-04, Royal Holloway, 1998.
[8] Tong Zhang. Leave-one-out bounds for kernel methods. Neural Computation,
15:1397?1437, 2003.
[9] Tong Zhang. Statistical analysis of some multi-category large margin classification
methods. Journal of Machine Learning Research, 5:1225?1251, 2004.
[10] Tong Zhang. Statistical behavior and consistency of classification methods based on
convex risk minimization. The Annals of Statitics, 32:56?85, 2004. with discussion.
.
| 2657 |@word pw:2 pick:1 tr:1 com:1 parsing:1 ronald:1 limp:2 selected:1 xk:21 completeness:1 provides:1 zhang:4 daphne:1 height:1 become:1 consists:2 expected:5 behavior:4 p1:1 nonseparable:1 multi:18 relying:1 decreasing:1 increasing:1 becomes:1 bounded:3 notation:2 underlying:1 moreover:4 what:2 pto:2 nj:1 guarantee:1 exactly:1 classifier:3 control:1 supc:7 negligible:1 local:1 limit:2 mistake:1 approximately:1 plus:1 studied:1 examined:1 practical:2 satisfactorily:1 practice:4 significantly:1 operator:4 risk:1 www:1 equivalent:1 map:1 center:1 convex:6 survey:1 simplicity:2 m2:7 rule:2 regarded:3 deriving:1 stability:2 handle:1 fx:29 annals:1 mcollins:1 programming:1 associate:1 ep:2 taskar:1 removed:3 yk:32 intuition:1 dynamic:1 weakly:1 solving:1 depend:2 tight:1 h0:7 quite:1 heuristic:1 larger:1 valued:1 widely:1 favor:2 rosenfeld:1 transform:1 sequence:4 advantage:2 propose:1 subtracting:1 product:1 gen:60 achieve:1 kh:8 olkopf:1 webpage:1 satellite:1 leave:1 ben:1 derive:2 measured:1 keywords:2 solves:2 come:2 implies:5 drawback:1 correct:1 generalization:32 investigation:1 summation:9 pl:4 hold:1 sufficiently:1 exp:1 k2h:4 great:1 algorithmic:1 lawrence:1 claim:1 achieves:1 radiance:1 purpose:2 estimation:1 bridge:1 successfully:1 establishes:1 minimization:1 mit:2 clearly:2 gaussian:5 pn:7 publication:1 corollary:2 derived:2 focus:2 flaw:1 dependent:1 typically:1 koller:1 interested:3 arg:6 classification:23 issue:2 html:1 smoothing:2 special:4 kw:4 koby:1 report:1 serious:1 divergence:1 replaced:3 multiply:1 introduces:1 behind:1 phc:1 bregman:1 nowadays:1 penalizes:4 desired:1 theoretical:4 earlier:1 soft:1 statitics:1 subset:1 predictor:2 dependency:2 supx:2 lee:1 michael:1 choose:3 possibly:1 iwpt:1 american:1 style:1 leading:1 li:12 potential:2 wk:47 includes:1 rockafellar:1 matter:1 ranking:1 later:1 view:2 try:1 kwk:7 sup:13 start:1 bayes:2 carlos:1 annotation:2 ni:1 efficiently:1 maxc:1 canceling:1 sebastian:1 definition:1 associated:1 proof:8 mi:2 newly:1 popular:2 knowledge:2 stanley:1 hilbert:2 routine:1 formulation:16 done:1 just:1 hand:10 sketch:4 parse:1 web:2 logistic:3 quality:1 behaved:1 believe:2 effect:1 contain:1 true:3 remedy:6 verify:1 regularization:3 equality:3 satisfactory:1 attractive:2 encourages:1 covering:1 yorktown:1 performs:1 superior:1 discussed:2 association:1 cambridge:1 ai:1 consistency:3 examp:1 i6:2 pointed:1 language:2 dot:1 pq:1 showed:2 belongs:1 inf:3 certain:1 inequality:4 watson:2 arbitrarily:1 inconsistency:1 yi:20 guestrin:1 multiple:4 desirable:3 reduces:1 technical:1 match:2 characterized:1 lin:1 variant:1 regression:3 kwk2h:6 redefinition:1 expectation:5 essentially:1 kernel:4 achieved:1 penalize:2 want:2 microarray:1 crucial:1 appropriately:2 sch:1 unlike:1 inconsistent:1 near:2 counting:1 easy:1 wahba:1 reduce:1 multiclass:1 penalty:1 speech:1 ignored:1 useful:1 clear:1 category:16 http:1 specifies:1 discrete:1 shall:1 drawn:2 penalizing:2 subgradient:2 nkwk:2 tzhang:1 decision:1 appendix:2 scaling:2 bound:30 replaces:1 oracle:1 infinity:1 aspect:1 extremely:1 min:4 separable:11 poor:1 ln:8 discus:1 loose:2 count:2 singer:1 end:3 generalizes:1 available:1 fxi:12 batch:1 denotes:1 include:2 cf:1 hinge:1 multicategory:1 yoram:1 analsysis:1 thrun:1 capacity:1 w0:3 me:1 argue:1 trivial:1 length:1 potentially:1 negative:1 implementation:1 unknown:1 markov:1 finite:3 reproducing:1 smoothed:1 arbitrary:1 princeton:2 pair:2 specified:1 sentence:3 dli:2 trans:1 able:2 suggested:1 kwkh:3 mismatch:1 max:7 including:1 royal:1 misclassification:3 suitable:2 natural:2 regularized:6 lk:6 hm:1 prior:3 asymptotic:1 loss:10 interesting:2 supy:2 consistent:2 editor:1 ibm:2 translation:1 penalized:1 last:2 free:1 english:1 side:10 perceptron:1 saul:1 taking:2 made:3 ple:1 kfx:3 approximate:1 bernhard:1 reveals:1 summing:2 discriminative:5 xi:20 search:3 zk:16 investigated:1 hc:3 separator:1 linearly:2 csd:1 bounding:1 motivation:1 referred:1 ny:1 tong:4 exponential:1 candidate:1 pe:2 watkins:1 formula:1 theorem:16 fxk:10 showing:1 jensen:1 symbol:1 svm:8 dl:1 exists:1 importance:1 margin:11 chen:1 entropy:13 ma:1 weston:1 conditional:4 replace:1 hard:2 included:1 specifically:1 tyrrell:1 lemma:4 called:1 e:12 holloway:1 support:7 people:1 latter:1 crammer:1 collins:1 audio:1 ex:2 |
1,825 | 2,658 | Generative Affine Localisation and Tracking
John Winn
Andrew Blake
Microsoft Research Cambridge
Roger Needham Building
7 J. J. Thomson Avenue
Cambridge CB3 0FB, U.K
http://research.microsoft.com/mlp
Abstract
We present an extension to the Jojic and Frey (2001) layered sprite model
which allows for layers to undergo affine transformations. This extension
allows for affine object pose to be inferred whilst simultaneously learning the object shape and appearance. Learning is carried out by applying
an augmented variational inference algorithm which includes a global
search over a discretised transform space followed by a local optimisation. To aid correct convergence, we use bottom-up cues to restrict the
space of possible affine transformations. We present results on a number
of video sequences and show how the model can be extended to track an
object whose appearance changes throughout the sequence.
1
Introduction
Generative models provide a powerful and intuitive way to analyse images or video sequences. Because such models directly represent the process of image generation, it is
straightforward to incorporate prior knowledge about the imaging process and to interpret
results. Since the entire data set is modelled, generative models can give improved accuracy and reliability over feature-based approaches and they also allow for selection between
models using Bayesian model comparison. Finally, it is possible to sample from generative
models, for example, for the purposes of image or video editing.
One popular type of generative model represents images as a composition of layers [1]
where each layer corresponds to the appearance and shape of an individual object or the
background. If the generative model is expressed probabilistically, Bayesian learning and
inference techniques can then be applied to reverse the imaging process and infer the shape
and appearance of individual objects in an unsupervised fashion [2].
The difficulty with generative models is how to apply Bayesian inference efficiently. In a
layered model, inference involves localising the pose of the layers in each image, which
is hard because of the large space of possible object transformations that needs to be explored. Previously, this has been dealt with by imposing restrictions on the space of object
transformations, such as allowing only similarity transformations [3]. Alternatively, if the
images are known to belong to a video sequence, tracking constraints can be used to focus the search on a small area of transformation space consistent with a dynamic model of
object motion [4]. However, even in a video sequence, this technique relies on the object
remaining in frame and moving relatively slowly.
In this paper, we extend the work of [3] and present an approach to object localisation which
allows objects to undergo planar affine transformations and works well both in the frames of
a video sequence and in unordered sets of images. A two-layer generative model is defined
and inference performed using a factorised variational approximation, including a global
search over a discretised transform space followed by a local optimisation using conjugate
gradients. Additionally, we exploit bottom-up cues to constrain the space of transforms
being explored. Finally, we extend our generative model to allow the object appearance in
one image to depend on its appearance in the previous one. Tracking appearance in this
way gives improved performance for objects whose appearance changes slowly over time
(e.g. objects undergoing non-planar rotation).
If the images are not frames of a video, or the object is out-of-frame or occluded in the
previous image, then the system automatically reverts to using a learned foreground appearance model.
2
The generative image model
This section describes the generative image model, which is illustrated in the Bayesian
network of Figure 1. This model consists of two layers, a foreground layer containing a
single object and a background layer.
We denote our image set as {x1 , . . . , xN }, where xi is a vector of the pixel intensities in the
ith image. The background layer is assumed to be stationary and so its appearance vector b
is set to be the same size as the image. A mask mi has binary elements that indicate which
b
f
T
m
x
N
Figure 1: The Bayesian network for the generative image model. The rounded rectangle is a
plate, indicating that there are N copies of each contained node (one for each image). Common
to all images are the background b, foreground object appearance f and mask prior ?. An affine
transform T gives the position and pose of the object in each image. The binary mask m defines
the area of support of the foreground object and has a prior given by a transformed ?. The observed
image x is generated by adding noise ? separately to the transformed foreground appearance and the
background and composing them together using the mask. For illustration, the images underneath
each node of the graph represent the inferred value of that node given a data set of hand images. A
priori, the appearance and mask of the object are not known.
pixels of the ith image are foreground. The mask is set to be slightly larger than the image
to allow the foreground object to overlap the edge of the image.
The foreground layer is represented by an appearance image vector f and a prior over its
mask ?, both of which are to be inferred from the image set. The elements of ? are real
numbers in the range [0, 1] which indicate the probability that the corresponding mask
pixels are on, as suggested in [5]. The object appearance and mask prior are defined in a
canonical, normalised pose; the actual position and pose of the object in the ith image is
given by an affine transformation Ti . With our images in vector form, we can consider a
transformation T to be a sparse matrix where the jth row defines the linear interpolation
of pixels that gives the jth pixel in the transformed image. For example, a translation of an
integer number of pixels is represented as a matrix whose entries Tjk are 1 if the translation
of location k in the source image is location j in the destination image, and 0 otherwise.
Hence, the transformed foreground appearance is given by Tf and the transformed mask
prior by T?. Given the transformed mask prior, the conditional distribution for the kth
mask pixel is
P (mk = 1 | ?, T) = (T?)k .
(1)
The observed image x is generated by a composition of the transformed foreground appearance and the background plus some noise. The conditional distribution for the kth image
pixel is given by
P (xk | b, f , m, T, ?) = N (xk | (Tf )k , ?f?1 )mk N (xk | bk , ?b?1 )1?mk
(2)
where ? = (?f , ?b ) are the noise precisions for the foreground layer and the background
layer respectively. The elements of both b and f are given broad Gaussian priors and the
prior on each ? is a broad Gamma distribution. The prior on each element of ? is a Beta
distribution.
3
Factorised variational inference
Given the above model and a set of images {x1 , . . . , xN }, the inference task is to learn
a posterior distribution over all other variables including the background, the foreground
appearance and mask prior, the transformation and mask for each image and the noise
precisions. Direct application of Bayes?s theorem is intractable because this would require
integrating over all unobserved variables. Instead, we turn to the approximate inference
technique of variational inference [6].
Variational inference involves defining a factorised variational distribution Q and then optimising to minimise the Kullback-Leibler divergence between Q and the true posterior
distribution. The motivation behind this methodology is that we expect the posterior to be
unimodal and tightly peaked and so it can be well approximated by a separable distribution.
In this paper, we choose our variational distribution to be factorised with respect to each element of b, f , m and ? and also with respect to ?f and ?b . The factor of Q corresponding
to each of these variables has the same form as the prior over that variable. For example,
the factor for the kth element of ? is a Beta distribution Q(?k ) = ?(?k | a0 , b0 ). The choice
of approximation to the posterior over the affine transform Q(T) is a more complex one,
and will be discussed below.
The optimisation of the Q distribution is achieved by firstly initialising the parameters of all
factors and then iteratively updating each factor in turn so as to minimise the KL divergence
whilst keeping all other factors fixed. If we define H to be the set of all hidden variables,
then the factor for the ith member Hi is updated using
log Q(Hi ) = hlog P ({x1 , . . . , xN }, H)i?Q(Hi ) + const.
(3)
f
b
b
b2
f
f2
? f T (m * x)
?
1
2
? fT m
T
log ?
x
? 12 (1 ? m) * ( x ? b) 2
?
1
2
m * ( x ? Tf )
T log ?
T log(1 ? ? )
? f m* x
? 12 ? f m
? b (1 ? m) * x
? 12 ? b (1 ? m)
1
2
1
2
2
T ?1m
1 ? T ?1m
?1
Tf
Tf 2
?
log ?
log(1 ? ? )
?1
m
1? m
m
1? m
m
log ? b ? ? b ( x ? b) 2
log ? f ? ? f ( x ? Tf ) 2
N
Figure 2: The messages passed when VMP is applied to the generative model. The messages to
or from T are not shown (see text). Where a message is shown as leaving the N plate, the destination
node receives a set of N messages, one from each copy of the nodes within the plate. Where a
message is shown entering the N plate, the message is sent to all copies of the destination node. All
expectations are with respect to the variational distribution Q.
where h.i?Q(Hi ) means the expectation under the distribution given by the product of all
factors of Q except Q(Hi ).
When the model is a Bayesian network, this optimisation procedure can be carried out
in a modular fashion by applying Variational Message Passing (VMP) [7, 8]. Using VMP
makes it very much simpler and quicker to extend, modify, combine or compare probabilistic models; it gives the same results as applying factorised variational inference by hand and
places no additional constraints on the form of the model. In VMP, messages consisting
of vectors of real numbers are sent to each node from its parent and children in the graph.
In our model, the messages to and from all nodes (except T) are shown in Figure 2. By
expressing each variational factor as an exponential family distribution, the ?natural parameter vector? [8] of that distribution can be optimised using (3) by adding messages received
at the corresponding node. For example, if the prior over b is N (b | ?, ? ?1 ), the parameter
vector of the factor Q(b) = N (b | ?0 , ? 0?1 ) is updated from the messages received at b
using
prior
natural param. vector
z
}| {
?0 ? 0
? 21 ? 0
=
z
received messages
z
}| { N
}|
{
X h? (1 ? m ) ? x i
??
bi
i
i
+
.
? 12 ?
? 12 h?bi (1 ? mi )i
(4)
i=1
The form of the natural parameter vector varies for different exponential family distributions (Gaussian, Gamma, Beta, discrete . . . ) but the update equation remains the same.
Following this update, the message being sent from b is recomputed to reflect the new parameters of Q(b). For details of the derivation this update equation and how to determine
VMP messages for a given model, see [8].
Where a set of similar messages are sent corresponding to the pixels of an image, it is
convenient to think instead of a single message where each element is itself an image. It
is efficient to structure the implementation in this way because message computation and
parameter updates can then be carried out using block operations on entire images.
4
Learning the object transformation
Following [3], we decompose the layer transformation into a product of transformations
and define a variational distribution that is separable over each. To allow for affine transformations, we choose to decompose T into three transformations applied sequentially,
T = Txy Trs Ta .
(5)
In this expression, Txy is a two-dimensional translation belonging to a finite set of translations Txy . Similarly, Trs is a rotation and uniform scaling and so the space of transforms
is also two-dimensional and is discretised to form a finite set Trs . The third transformation
Ta is a freeform (non-discretised) affine transform. The variational distribution over the
combined transform T is given by
Q(T) = Q(Txy )Q(Trs )Q(Ta ).
(6)
Because Txy and Trs are discretised, Q(Txy ) and Q(Trs ) are defined to be discrete distributions. We can apply (3) to determine the update equations for these distributions,
= hmi . (Txy hTrs Ta log ?i) + h1 ? mi . (Txy hTrs Ta log(1 ? ?)i)
(7)
+ ?f hmi . x ? Txy hTrs Ta f i ? 12 Txy Trs Ta f 2 + z xy
?1
?1
log Q(Trs ) = Txy m . (Trs hTa log ?i) + Txy (1 ? m) . (Trs hTa log(1 ? ?)i)
?1
1
+ ?f T?1
Trs Ta f 2 + z rs (8)
xy (m ? x) . (Trs hTa f i) ? 2 ?f Txy m
log Q(Txy )
where z xy and z rs are constants which can be found by normalisation.
As described in [3], the evaluation of (7) and (8) for all Txy ? Txy and all Trs ? Trs
can be carried out efficiently using Fast Fourier Transforms in either Cartesian or log-polar
co-ordinate systems. The use of FFTs allows us to make both Txy and Trs large: we set Txy
to contain all translations of a whole number of pixels and Trs to contain 360 rotations (at
1? intervals) and 50 scalings (where each scaling represents a ? 1.5% increase in length
scale). FFTs can be used within the VMP framework as both (7) and (8) involve quantities
that are contained in messages to T (see Figure 2).
Finally, we define the variational distribution over Ta to be a delta function,
Q(Ta ) = ?(Ta ? T?a ).
(9)
Unlike all the other variational factors, this cannot be optimised analytically. To minimise
the KL divergence, we need to find the value of T?a that maximises
?1 ?1
?1
?
?
Fa = T?1
rs Txy m . (Ta hlog ?i) + Trs Txy (1 ? m) . (Ta hlog(1 ? ?)i)
?1 ?1
?1
T?a f 2 .
(10)
+ ?f Trs Txy (m ? x) . (T?a hf i) ? 12 ?f T?1
rs Txy m
This local maximisation is achieved efficiently by using a trust-region Newton method. The
assumption is that the search through Txy and Trs has located the correct posterior mode in
transform space and that it is only necessary to use gradient methods to find the peak of that
mode. This assumption appeared valid for the image sequences used in our experiments,
even when the transformation of the foreground layer was not well approximated by a
similarity transform alone.
Inference in this model is made harder due to an inherent non-identifiability problem. The
pose of the learned appearance and mask prior is undefined and so applying a transform to f
and ? and the inverse of the transform to each Ti results in an unchanged joint distribution.
When applying a variational technique, such non-identifiability leads to many more local
minima in the KL divergence. We partially resolve this issue by adding a constraint to this
model that the expected mask h?i is centred, so that its centre of gravity is in the middle
of the latent image. This constraint is applied by shifting the parameters of Q(?) directly
following each update (and also shifting Q(f ) and each Q(T) appropriately).
Background
Example frame #1
Foreground #1
Normalised frame #1
Object appearance & mask
Example frame #2
Foreground #2
Normalised frame #2
Figure 3: Tracking a hand undergoing extreme affine transformation. The first column shows
the learned background and masked object appearance. The second and third columns contain two
frames from the sequence along with the foreground segmentation for each. The final column shows
each frame transformed by the inverse of the inferred object transform. In each image the red outline
surrounds the area where the transformed mask prior ? is greater than 0.5.
4.1
Using bottom-up information to improve inference
Given that ? is centred, we can significantly improve convergence by using bottom-up information about the translation of the object. For example, the inferred mask m i for each
frame is very informative about the location of the object in that frame. Using sufficient
data, we could learn a conditional model P (Txy | hmi i) and bound Txy by only considering translations with non-negligible posterior mass under this conditional model. Instead,
we use a conservative, hand-constructed bound on Txy based on the assumption that, during inference, the most probable mask under Q(mi ) consists of a (noisy) subset of the true
mask pixels. Suppose the true mask contains M non-zero pixels with second moment of
area IM and the current most probable mask contains V non-zero pixels (V ? M ) with
second moment of area IV . A bound on c, the position of the centre of the inferred mask
relative to the centre of true mask, is given by
diag(ccT ) ? (M ? V ) diag(IM /V ? IV /M ).
(11)
We can gain a conservative estimate of M and IM by using the maximum values of V
and IV across all frames, multiplied by a constant ? ? 1.2. The bound is deliberately
constructed to be conservative; its purpose is to discard settings of Txy that have negligible
probability under the model and so avoid local minima in the variational optimisation. The
bound is updated at each iteration and applied by setting Q(Txy ) = 0 for values of Txy
outside the bound. Q(Txy ) is then re-normalised.
The use of this bound on Txy is intended as a very simple example of incorporating bottomup information to improve inference within a generative model. In future work, we intend
to investigate using more informative bottom-up cues, such as optical flow or tracked interest points, to propose probable transformations within this model. Incorporating such
proposals or bounds into a variational inference framework both speeds convergence and
helps avoid local minima.
5
Experimental results
We present results on two video sequences. The first is of a hand rotating both parallel
to the image plane and around its own axis, whilst also translating in three dimensions.
The sequence consists of 59 greyscale frames, each of size 160 ? 120 pixels (excluding
the border). Our Matlab implementation took about a minute per frame to analyse the se-
Appearance & mask
Example frame #1
Foreground #1
Example frame #2
Foreground #2
Figure 4: Affine tracking of a semi-transparent object.
Appearance & mask
First frame
Foreground
Last frame
Foreground
Figure 5: Tracking an object with changing appearance. A person is tracked throughout a sequence despite their appearance changing dramatically from between first and last frames. The blue
outline shows the inferred mask m which differs slightly from ? due to the object changing shape.
quence, over half of which was spent on the conjugate gradient optimisation step. Figure 3
shows the expected values of the background and foreground layers under the optimised
variational distribution, along with foreground segmentation results for two frames of the
sequence. The right hand column gives another indication of the accuracy of the inferred
transformations by applying the inverse transformation to the entire frame and showing that
the hand then has a consistent normalised position and size. In a video of the hand showing
the tracked outline,1 the outline appears to move smoothly and follow the hand with a high
degree of accuracy, despite the system not using any temporal constraints.
Results for a second sequence showing a cyclist are given in Figure 4. Although the cyclist
and her shadow are tracked correctly, the learned appearance is slightly inaccurate as the
system is unable to capture the perspective foreshortening of the bicycle. This could be
corrected by allowing Ta to include projective transformations.
6
Tracking objects with changing appearance
The model described so far makes the assumption that the appearance of the object does
not change significantly from frame to frame. If the set of images are actually frames from
a video, we can model objects whose appearance changes slowly by allowing the model
to use the object appearance in the previous frame as the basis for its appearance in the
current frame. However, we may not know if the images are video frames and, even if we
do, the object may be occluded or out-of-frame in the previous image. We can cope with
this uncertainty by inferring automatically whether to use the previous frame or the learned
appearance f . Switching between two methods in this way is similar to [9].
The model is extended by introducing a binary variable si for each frame and define a new
appearance variable gi = si f + (1 ? si )T?1
i?1 xi?1 . Hence gi either equals the foreground
appearance f (if si = 1) or the transform-normalised previous frame (if si = 0). For the
first frame, we fix s1 = 1. We then replace f with gi in (2) and then apply VMP within the
resulting Bayesian network.
The extended model is able to track an object even when its appearance changes significantly throughout the image sequence (see Figure 5). The binary variable s i is found to
have an expected value ? 0 for all frames (except the first). Using the tracked appearance
1
Videos of results are available from http://johnwinn.org/Research/affine.
allows the foreground segmentation of each frame to be accurate even though the object is
poorly modelled by the inferred appearance image. If we introduce an abrupt change into
the sequence, for example by reversing the second half of the sequence, hs i i is found to be
? 1 for the frame following the change. In other words, the system has detected not to use
the previous frame at this point, but to revert to using the latent appearance image f .
7
Discussion
We have proposed a method for localising an object undergoing affine transform whilst
simultaneously learning its shape and appearance. This power of this method has been
demonstrated by tracking moving objects in several real videos, including where the appearance of the object changes significantly from start to end. The system makes no assumptions about the speed of motion of the object, requires no special initialisation and is
robust to the object being temporarily occluded or moving out of frame.
A natural extension to this work is to allow multiple layers, with each layer having its own
latent shape and appearance and set of affine transformations. Unfortunately, as the number of latent variables increases, the inference problem becomes correspondingly harder
and an exhaustive search becomes less practical. Instead, we are investigating performing inference in a simpler model where a subset of the variables have been approximately
marginalised out. The results of using this simpler model can then be used to guide inference in the full model. A further interesting addition to the model would be to allow layers
to be grouped into rigid or articulated three-dimensional objects.
Acknowledgments
The authors would like to thank Nebojsa Jojic for suggesting the use of a binary switch
variable for tracking and Tom Minka for helpful discussions.
References
[1] J. Y. A. Wang and E. H. Adelson. Representing moving images with layers. In IEEE Transactions
on Image Processing, volume 3, pages 625?638, 1994.
[2] N. Jojic and B. Frey. Learning flexible sprites in video layers. In Proc. of IEEE Conf. on
Computer Vision and Pattern Recognition, 2001.
[3] B. Frey and N. Jojic. Fast, large-scale transformation-invariant clustering. In Advances in Neural
Information Processing Systems 14, 2001.
[4] M. K. Titsias and C. K. I. Williams. Fast unsupervised greedy learning of multiple objects
and parts from video. 2004. To appear in Proc. Generative-Model Based Vision Workshop,
Washington DC, USA.
[5] C.K.I. Williams and M. K. Titsias. Greedy learning of multiple objects in images using robust
statistics and factorial learning. Neural Computation, 16(5):1039?1062, 2004.
[6] M. I. Jordan, Z. Ghahramani, T. S. Jaakkola, and L. K. Saul. An introduction to variational
methods for graphical models. In M. I. Jordan, editor, Learning in Graphical Models, pages
105?162. Kluwer, 1998.
[7] C. M. Bishop, J. M. Winn, and D. Spiegelhalter. VIBES: A variational inference engine for
Bayesian networks. In Advances in Neural Information Processing Systems, volume 15, 2002.
[8] J. M. Winn and C. M. Bishop. Variational Message Passing. 2004. To appear in Journal of
Machine Learning Research. Available from http://johnwinn.org.
[9] A. Jepson, D. Fleet, and T. El-Maraghi. Robust online appearance models for visual tracking. In
Proc. IEEE Conf. Computer Vision and Pattern Recognition, volume I, pages 415?422, 2001.
| 2658 |@word h:1 middle:1 r:4 harder:2 moment:2 contains:2 initialisation:1 current:2 com:1 si:5 john:1 informative:2 shape:6 update:6 stationary:1 generative:15 cue:3 alone:1 half:2 nebojsa:1 greedy:2 plane:1 xk:3 ith:4 node:9 location:3 firstly:1 simpler:3 org:2 along:2 constructed:2 direct:1 beta:3 consists:3 combine:1 introduce:1 mask:28 expected:3 cct:1 automatically:2 resolve:1 actual:1 param:1 considering:1 becomes:2 mass:1 whilst:4 unobserved:1 transformation:24 temporal:1 ti:2 gravity:1 appear:2 negligible:2 frey:3 local:6 modify:1 switching:1 despite:2 optimised:3 interpolation:1 approximately:1 plus:1 co:1 projective:1 range:1 bi:2 practical:1 acknowledgment:1 maximisation:1 block:1 differs:1 procedure:1 area:5 significantly:4 convenient:1 word:1 integrating:1 cannot:1 layered:2 selection:1 applying:6 restriction:1 demonstrated:1 straightforward:1 williams:2 txy:31 abrupt:1 updated:3 suppose:1 element:7 approximated:2 recognition:2 updating:1 located:1 bottom:5 observed:2 ft:1 quicker:1 wang:1 capture:1 region:1 occluded:3 dynamic:1 depend:1 titsias:2 f2:1 basis:1 joint:1 represented:2 derivation:1 articulated:1 revert:1 fast:3 detected:1 outside:1 exhaustive:1 whose:4 modular:1 larger:1 otherwise:1 statistic:1 gi:3 think:1 transform:13 analyse:2 itself:1 final:1 noisy:1 online:1 sequence:16 indication:1 took:1 propose:1 product:2 poorly:1 intuitive:1 convergence:3 parent:1 object:47 help:1 spent:1 andrew:1 pose:6 received:3 b0:1 involves:2 indicate:2 shadow:1 correct:2 translating:1 require:1 transparent:1 fix:1 maraghi:1 decompose:2 probable:3 im:3 extension:3 around:1 blake:1 bicycle:1 tjk:1 purpose:2 polar:1 proc:3 grouped:1 tf:6 gaussian:2 avoid:2 jaakkola:1 probabilistically:1 focus:1 quence:1 underneath:1 helpful:1 inference:20 rigid:1 el:1 inaccurate:1 entire:3 a0:1 hidden:1 her:1 transformed:9 pixel:14 issue:1 flexible:1 priori:1 special:1 equal:1 having:1 washington:1 optimising:1 represents:2 broad:2 adelson:1 unsupervised:2 vmp:7 foreground:24 peaked:1 future:1 inherent:1 foreshortening:1 simultaneously:2 gamma:2 divergence:4 individual:2 tightly:1 intended:1 consisting:1 microsoft:2 mlp:1 message:19 normalisation:1 investigate:1 localisation:2 interest:1 evaluation:1 extreme:1 undefined:1 behind:1 accurate:1 edge:1 necessary:1 xy:3 iv:3 rotating:1 re:1 mk:3 column:4 introducing:1 entry:1 subset:2 uniform:1 masked:1 varies:1 combined:1 person:1 peak:1 destination:3 probabilistic:1 rounded:1 together:1 reflect:1 containing:1 choose:2 slowly:3 conf:2 suggesting:1 factorised:5 centred:2 unordered:1 b2:1 includes:1 performed:1 h1:1 red:1 start:1 bayes:1 hf:1 parallel:1 identifiability:2 accuracy:3 efficiently:3 dealt:1 modelled:2 bayesian:8 minka:1 mi:4 gain:1 popular:1 knowledge:1 segmentation:3 actually:1 appears:1 ta:14 follow:1 planar:2 methodology:1 improved:2 tom:1 editing:1 though:1 roger:1 hand:9 receives:1 trust:1 defines:2 mode:2 usa:1 building:1 contain:3 true:4 deliberately:1 hence:2 analytically:1 jojic:4 entering:1 leibler:1 iteratively:1 illustrated:1 during:1 plate:4 outline:4 thomson:1 motion:2 image:52 variational:22 common:1 rotation:3 tracked:5 volume:3 extend:3 discussed:1 belong:1 kluwer:1 interpret:1 expressing:1 composition:2 cambridge:2 imposing:1 surround:1 similarly:1 centre:3 reliability:1 moving:4 similarity:2 posterior:6 own:2 perspective:1 reverse:1 discard:1 binary:5 minimum:3 additional:1 greater:1 determine:2 semi:1 multiple:3 unimodal:1 full:1 infer:1 optimisation:6 expectation:2 vision:3 iteration:1 represent:2 achieved:2 proposal:1 background:11 addition:1 separately:1 winn:3 interval:1 source:1 leaving:1 appropriately:1 unlike:1 undergo:2 sent:4 member:1 flow:1 jordan:2 integer:1 switch:1 restrict:1 avenue:1 minimise:3 fleet:1 whether:1 expression:1 passed:1 sprite:2 passing:2 matlab:1 dramatically:1 se:1 involve:1 factorial:1 transforms:3 http:3 canonical:1 delta:1 track:2 per:1 correctly:1 blue:1 discrete:2 recomputed:1 cb3:1 changing:4 rectangle:1 imaging:2 graph:2 hta:3 inverse:3 powerful:1 uncertainty:1 place:1 throughout:3 family:2 initialising:1 scaling:3 layer:20 hi:5 bound:8 followed:2 trs:19 constraint:5 constrain:1 fourier:1 speed:2 performing:1 separable:2 optical:1 relatively:1 conjugate:2 belonging:1 describes:1 slightly:3 across:1 s1:1 invariant:1 equation:3 previously:1 remains:1 turn:2 know:1 end:1 available:2 operation:1 multiplied:1 apply:3 localising:2 remaining:1 include:1 clustering:1 graphical:2 newton:1 const:1 exploit:1 ghahramani:1 unchanged:1 move:1 intend:1 quantity:1 fa:1 gradient:3 kth:3 unable:1 thank:1 length:1 illustration:1 unfortunately:1 hlog:3 greyscale:1 implementation:2 allowing:3 maximises:1 finite:2 defining:1 extended:3 excluding:1 frame:38 dc:1 intensity:1 inferred:9 ordinate:1 bk:1 discretised:5 kl:3 engine:1 learned:5 able:1 suggested:1 below:1 pattern:2 appeared:1 reverts:1 including:3 video:15 shifting:2 power:1 overlap:1 difficulty:1 natural:4 marginalised:1 representing:1 improve:3 spiegelhalter:1 axis:1 carried:4 text:1 prior:16 relative:1 expect:1 generation:1 interesting:1 degree:1 affine:15 sufficient:1 consistent:2 editor:1 translation:7 row:1 last:2 copy:3 keeping:1 jth:2 guide:1 allow:6 normalised:6 saul:1 correspondingly:1 sparse:1 dimension:1 xn:3 valid:1 fb:1 author:1 made:1 far:1 cope:1 transaction:1 approximate:1 kullback:1 global:2 sequentially:1 investigating:1 assumed:1 xi:2 alternatively:1 bottomup:1 search:5 latent:4 additionally:1 learn:2 robust:3 composing:1 hmi:3 cyclist:2 complex:1 diag:2 jepson:1 motivation:1 noise:4 whole:1 border:1 child:1 x1:3 augmented:1 fashion:2 aid:1 precision:2 position:4 inferring:1 exponential:2 third:2 ffts:2 theorem:1 minute:1 bishop:2 showing:3 undergoing:3 explored:2 intractable:1 incorporating:2 workshop:1 adding:3 cartesian:1 vibe:1 smoothly:1 appearance:42 visual:1 expressed:1 contained:2 tracking:10 partially:1 temporarily:1 corresponds:1 relies:1 conditional:4 replace:1 change:8 hard:1 except:3 corrected:1 reversing:1 conservative:3 experimental:1 indicating:1 support:1 incorporate:1 |
1,826 | 2,659 | Learning syntactic patterns for automatic
hypernym discovery
Rion Snow
Daniel Jurafsky
Andrew Y. Ng
Computer Science Department
Stanford University
Stanford, CA 94305
Linguistics Department
Stanford University
Stanford, CA 94305
Computer Science Department
Stanford University
Stanford, CA 94305
[email protected]
[email protected]
[email protected]
Abstract
Semantic taxonomies such as WordNet provide a rich source of knowledge for natural language processing applications, but are expensive to
build, maintain, and extend. Motivated by the problem of automatically
constructing and extending such taxonomies, in this paper we present a
new algorithm for automatically learning hypernym (is-a) relations from
text. Our method generalizes earlier work that had relied on using small
numbers of hand-crafted regular expression patterns to identify hypernym pairs. Using ?dependency path? features extracted from parse trees,
we introduce a general-purpose formalization and generalization of these
patterns. Given a training set of text containing known hypernym pairs,
our algorithm automatically extracts useful dependency paths and applies
them to new corpora to identify novel pairs. On our evaluation task (determining whether two nouns in a news article participate in a hypernym
relationship), our automatically extracted database of hypernyms attains
both higher precision and higher recall than WordNet.
1
Introduction
Semantic taxonomies and thesauri such as WordNet [5] are a key source of knowledge for
natural language processing applications, and provide structured information about semantic relations between words. Building such taxonomies, however, is an extremely slow
and labor-intensive process. Further, semantic taxonomies are invariably limited in scope
and domain, and the high cost of extending or customizing them for an application has
often limited their usefulness. Consequently, there has been significant recent interest in
finding methods for automatically learning taxonomic relations and constructing semantic
hierarchies. [1, 2, 3, 4, 6, 8, 9, 13, 15, 17, 18, 19, 20, 21]
In this paper, we build an automatic classifier for the hypernym/hyponym relation. A noun
X is a hyponym of a noun Y if X is a subtype or instance of Y. Thus ?Shakespeare? is a
hyponym of ?author? (and conversely ?author? is a hypernym of ?Shakespeare?), ?dog? is
a hyponym of ?canine?, ?desk? is a hyponym of ?furniture?, and so on.
Much of the previous work on automatic semantic classification of words has been based
on a key insight first articulated by Hearst [8], that the presence of certain ?lexico-syntactic
patterns? can indicate a particular semantic relationship between two nouns. Hearst noticed
that, for example, linking two noun phrases (NPs) via the constructions ?Such NP Y as
NP X ?, or ?NP X and other NP Y ?, often implies that NP X is a hyponym of NP Y , i.e., that
NP X is a kind of NP Y . Since then, several researchers have used a small number (typically
less than ten) of hand-crafted patterns like these to try to automatically label such semantic
relations [1, 2, 6, 13, 17, 18]. While these patterns have been successful at identifying some
examples of relationships like hypernymy, this method of lexicon construction is tedious
and severely limited by the small number of patterns typically employed.
Our goal is to use machine learning to automatically replace this hand-built knowledge.
We first use examples of known hypernym pairs to automatically identify large numbers of
useful lexico-syntactic patterns, and then combine these patterns using a supervised learning algorithm to obtain a high accuracy hypernym classifier. More precisely, our approach
is as follows:
1. Training:
(a) Collect noun pairs from corpora, identifying examples of hypernym pairs
(pairs of nouns in a hypernym/hyponym relation) using WordNet.
(b) For each noun pair, collect sentences in which both nouns occur.
(c) Parse the sentences, and automatically extract patterns from the parse tree.
(d) Train a hypernym classifier based on these features.
2. Test:
(a) Given a pair of nouns in the test set, extract features and use the classifier to
determine if the noun pair is in the hypernym/hyponym relation or not.
The rest of the paper is structured as follows. Section 2 introduces our method for automatically discovering patterns indicative of hypernymy. Section 3 then describes the setup of
our experiments. In Section 4 we analyze our feature space, and in Section 5 we describe
a classifier using these features that achieves high accuracy on the task of hypernym identification. Section 6 shows how this classifier can be improved by adding a new source of
knowledge, coordinate terms.
2
Representing lexico-syntactic patterns with dependency paths
The first goal of our work is to automatically identify lexico-syntactic patterns indicative
of hypernymy. In order to do this, we need a representation space for expressing these
patterns. We propose the use of dependency paths as a general-purpose formalization of
the space of lexico-syntactic patterns. Dependency paths have been used successfully in
the past to represent lexico-syntactic relations suitable for semantic processing [11].
A dependency parser produces a dependency tree that represents the syntactic relations
between words by a list of edge tuples of the form:
(word1 ,CATEGORY 1 :RELATION:CATEGORY 2 , word2 ). In this formulation each word is
the stemmed form of the word or multi-word phrase (so that ?authors? becomes ?author?),
and corresponds to a specific node in the dependency tree; each category is the part of
speech label of the corresponding word (e.g., N for noun or P REP for preposition); and
the relation is the directed syntactic relationship exhibited between word1 and word2
(e.g., OBJ for object, MOD for modifier, or CONJ for conjunction), and corresponds to a
specific link in the tree. We may then define our space of lexico-syntactic patterns to be all
shortest paths of four links or less between any two nouns in a dependency tree. Figure 1
shows the partial dependency tree for the sentence fragment ?...such authors as Herrick and
Shakespeare? generated by the broad-coverage dependency parser MINIPAR [10].
...
-N:pre:PreDet
authors
-N:mod:Prep
-N:punc:U
such
-Prep:pcomp-n:N
as
Herrick
and
-N:conj:N
-Prep:pcomp-n:N
Shakespeare
Figure 1: MINIPAR dependency tree example with transform
NP X and other NP Y :
NP X or other NP Y :
NP Y such as NP X :
Such NP Y as NP X :
NP Y including NP X :
NP Y , especially NP X :
(and,U: PUNC :N),-N: CONJ :N, (other,A: MOD :N)
(or,U: PUNC :N),-N: CONJ :N, (other,A: MOD :N)
N: PCOMP - N :P REP,such as,such as,P REP : MOD :N
N: PCOMP - N :P REP,as,as,P REP : MOD :N,(such,P RE D ET: PRE :N)
N: OBJ :V,include,include,V: I :C,dummy node,dummy node,C: REL :N
-N: APPO :N,(especially,A: APPO - MOD :N)
Table 1: Dependency path representations of Hearst?s patterns
We then remove the original nouns in the noun pair to create a more general pattern. Each
dependency path may then be presented as an ordered list of dependency tuples. We extend
this basic MINIPAR representation in two ways: first, we wish to capture the fact that certain function words like ?such? (in ?such NP as NP?) or ?other? (in ?NP and other NP?) are
important parts of lexico-syntactic patterns. We implement this by adding optional ?satellite links? to each shortest path, i.e., single links not already contained in the dependency
path added on either side of each noun. Second, we capitalize on the distributive nature of
the syntactic conjunction relation (nouns linked by ?and? or ?or?, or in comma-separated
lists) by distributing dependency links across such conjunctions. For example, in the simple 2-member conjunction chain of ?Herrick? and ?Shakespeare? in Figure 1, we add the
entrance link ?as, -P REP : PCOMP - N :N? to the single element ?Shakespeare? (as a dotted
line in the figure). Our extended dependency notation is able to capture the power of the
hand-engineered patterns described in the literature. Table 1 shows the six patterns used in
[1, 2, 8] and their corresponding dependency path formalizations.
3
Experimental paradigm
Our goal is to build a classifier which, when given an ordered pair of nouns, makes the
binary decision of whether the nouns are related by hypernymy.
All of our experiments are based on a corpus of over 6 million newswire sentences.1 We
first parsed each of the sentences in the corpus using MINIPAR. We extract every pair of
nouns from each sentence.
752,311 of the resulting unique noun pairs were labeled as Known Hypernym or Known
Non-Hypernym using WordNet.2 A noun pair (ni , nj ) is labeled Known Hypernym if nj
is an ancestor of the first sense of ni in the WordNet hypernym taxonomy, and if the only
?frequently-used?3 sense of each noun is the first noun sense listed in WordNet. Note that
nj is considered a hypernym of ni regardless of how much higher in the hierarchy it is with
respect to ni . A noun pair may be assigned to the second set of Known Non-Hypernym
pairs if both nouns are contained within WordNet, but neither noun is an ancestor of the
other in the WordNet hypernym taxonomy for any senses of either noun. Of our collected
noun pairs, 14,387 were Known Hypernym pairs, and we assign the 737,924 most frequently occurring Known Non-Hypernym pairs to the second set; this number is selected
to preserve the roughly 1:50 ratio of hypernym-to-non-hypernym pairs observed in our
hand-labeled test set (discussed below).
We evaluated our binary classifiers in two ways. For both sets of evaluations, our classifier
was given a pair of nouns from an unseen sentence and had to make a hypernym vs. nonhypernym decision. In the first style of evaluation, we compared the performance of our
classifiers against the Known Hypernym versus Known Non-Hypernym labels assigned by
1
The corpus contains articles from the Associated Press, Wall Street Journal, and Los Angeles
Times, drawn from the T IPSTER 1, 2, 3, and T REC 5 corpora [7]. Our most recent experiments
(presented in Section 6) include articles from Wikipedia (a popular web encyclopedia), extracted
with the help of Tero Karvinen?s Tero-dump software.
2
We access WordNet 2.0 via Jason Rennie?s WordNet::QueryData interface.
3
A noun sense is determined to be ?frequently-used? if it occurs at least once in the sense-tagged
Brown Corpus Semantic Concordance files (as reported in the cntlist file distributed as part of
WordNet 2.0). This determination is made so as to reduce the number of false hypernym/hyponym
classifications due to highly polysemous nouns (nouns which have multiple meanings).
Hypernym Classifiers on WordNet-labeled dev set
1
Logistic Regression (Buckets)
Logistic Regression (Binary)
Hearst Patterns
And/Or Other Pattern
0.9
0.8
Precision
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
Figure 2: Hypernym pre/re for all features
0
0.1
0.2
0.3
0.4
0.5
Recall
0.6
0.7
0.8
0.9
1
Figure 3: Hypernym classifiers
WordNet. This provides a metric for how well our classifiers do at ?recreating? WordNet?s
judgments.
For the second set of evaluations we hand-labeled a test set of 5,387 noun pairs from
randomly-selected paragraphs within our corpus (with part-of-speech labels assigned by
MINIPAR). The annotators were instructed to label each ordered noun pair as one of
?hyponym-to-hypernym?, ?hypernym-to-hyponym?, ?coordinate?, or ?unrelated? (the coordinate relation will be defined in Section 6). As expected, the vast majority of pairs
(5,122) were found to be unrelated by these measures; the rest were split evenly between
hypernym and coordinate pairs (134 and 131, resp.).
Interannotator agreement was obtained between four labelers (all native speakers of English) on a set of 511 noun pairs, and determined for each task according to the averaged
F-Score across all pairs of the four labelers. Agreement was 83% and 64% for the hypernym and coordinate term classification tasks, respectively.
4
Features: pattern discovery
Our first study focused on discovering which dependency paths might prove useful features
for our classifiers. We created a feature lexicon of 69,592 dependency paths, consisting of
every dependency path that occurred between at least five unique noun pairs in our corpus.
To evaluate these features, we constructed a binary classifier for each pattern, which simply
classifies a noun pair as hypernym/hyponym if and only if the specific pattern occurs at least
once for that noun pair. Figure 2 depicts the precision and recall of all such classifiers (with
recall at least .0015) on the WordNet-labeled data set.4 Using this formalism we have been
able to capture a wide variety of repeatable patterns between hypernym/hyponym noun
pairs; in particular, we have been able to rediscover the hand-designed patterns originally
proposed in [8] (the first five features, marked in red)5 , in addition to a number of new
patterns not previously discussed (of which four are marked as blue triangles in Figure
2 and listed in Table 2. This analysis gives a quantitative justification to Hearst?s initial
intuition as to the power of hand-selected patterns; nearly all of Hearst?s patterns are at the
high-performance boundary of precision and recall for individual features.
NP Y like NP X :
NP Y called NP X :
NP X is a NP Y :
NP X , a NP Y (appositive):
N: PCOMP - N :P REP,like,like,P REP : MOD :N
N: DESC :V,call,call,V: VREL :N
N: S :VBE,be,be,-VBE: PRED :N
N: APPO :N
Table 2: Dependency path representations of other high-scoring patterns
4
Redundant features consisting of an identical base path to an identified pattern but differing only
by an additional ?satellite link? are marked in Figure 2 by smaller versions of the same symbol.
5
We mark the single generalized ?conjunction other? pattern -N: CONJ :N, (other,A: MOD :N) to
represent both of Hearst?s original ?and other? and ?or other? patterns.
Best Logistic Regression (Buckets):
Best Logistic Regression (Binary):
Best Multinomial Naive Bayes:
Best Complement Naive Bayes:
Hearst Patterns:
?And/Or Other? Pattern:
0.3480
0.3200
0.3175
0.3024
0.1500
0.1170
Table 3: Average maximum F-scores for cross validation on WordNet-labeled training set
5
A hypernym-only classifier
Our first hypernym classifier is based on the intuition that unseen noun pairs are more
likely to be a hypernym pair if they occur in the test set with one or more lexico-syntactic
patterns found to be indicative of hypernymy. We record in our noun pair lexicon each
noun pair that occurs with at least five unique paths from our feature lexicon discussed in
the previous section. We then create a feature count vector for each such noun pair. Each
entry of the 69,592-dimension vector represents a particular dependency path, and contains
the total number of times that that path was the shortest path connecting that noun pair in
some dependency tree in our corpus. We thus define as our task the binary classification of
a noun pair as a hypernym pair based on its feature vector of dependency paths.
We use the WordNet-labeled Known Hypernym / Known Non-Hypernym training set defined in Section 3. We train a number of classifiers on this data set, including multinomial
Naive Bayes, complement Naive Bayes [16], and logistic regression. We perform model
selection using 10-fold cross validation on this training set, evaluating each model based
on its maximum F-Score averaged across all folds. The summary of average maximum
F-scores is presented in Table 3, and the precision/recall plot of our best models is presented in Figure 3. For comparison, we evaluate two simple classifiers based on past work
using only a handful of hand-engineered features; the first simply detects the presence of
at least one of Hearst?s patterns, arguably the previous best classifier consisting only of
lexico-syntactic patterns, and as implemented for hypernym discovery in [2]. The second
classifier consists of only the ?NP and/or other NP? subset of Hearst?s patterns, as used
in the automatic construction of a noun-labeled hypernym taxonomy in [1]. In our tests
we found greatest performance from a binary logistic regression model with 14 redundant
threshold buckets spaced at the exponentially increasing intervals {1, 2, 4, ...4096, 8192};
our resulting feature space consists of 974,288 distinct binary features. These buckets are
defined such that a feature corresponding to pattern p at threshold t will be activated by
a noun pair n if and only if p has been observed to occur as a shortest dependency path
between n at least t times.
Our classifier shows a dramatic improvement over previous classifiers; in particular, using
our best logistic regression classifier trained on newswire corpora, we observe a 132%
relative improvement of average maximum F-score over the classifier based on Hearst?s
patterns.
6
Using coordinate terms to improve hypernym classification
While our hypernym-only classifier performed better than previous classifiers based on
hand-built patterns, there is still much room for improvement. As [2] points out, one problem with pattern-based hypernym classifiers in general is that within-sentence hypernym
pattern information is quite sparse. Patterns are useful only for classifying noun pairs which
happen to occur in the same sentence; many hypernym/hyponym pairs may simply not occur in the same sentence in the corpus. For this reason [2], following [1] suggests relying
on a second source of knowledge: ?coordinate? relations between nouns. The WordNet
glossary defines coordinate terms as ?nouns or verbs that have the same hypernym?. Here
we treat the coordinate relation as a symmetric relation that exists between two nouns that
share at least one common ancestor in the hypernym taxonomy, and are therefore ?the
same kind of thing? at some level. Many methods exist for inferring that two nouns are
coordinate terms (a common subtask in automatic thesaurus induction). We expect that
Interannotator Agreement:
Distributional Similarity Vector Space Model:
Thresholded Conjunction Pattern Classifier:
Best WordNet Classifier:
0.6405
0.3327
0.2857
0.2630
Table 4: Summary of maximum F-scores on hand-labeled coordinate pairs
Coordinate term classifiers on hand-labeled test set
1
0.9
0.8
0.8
0.7
0.6
0.5
0.4
0.6
0.5
0.4
0.3
0.3
0.2
0.2
0.1
0
Interannotator Agreement
TREC+Wikipedia
TREC, Hybrid
TREC, Hypernym-only
WordNet Classifiers
Hearst Patterns
And/Or Other Pattern
0.9
Precision
Precision
0.7
Hypernym Classifiers on hand-labeled test set
1
Interannotator Agreement
Distributional Similarity
Conjunct Pattern
WordNet
0.1
0
0.1
0.2
0.3
0.4
0.5
Recall
0.6
0.7
Figure 4: Coordinate classifiers on
hand-labeled test set
0.8
0.9
1
0
0
0.1
0.2
0.3
0.4
0.5
Recall
0.6
0.7
0.8
0.9
Figure 5: Hypernym classifiers on
hand-labeled test set
using coordinate information will increase the recall of our hypernym classifier: if we are
confident that two nouns ni , nj are coordinate terms, and that nj is a hyponym of nk , we
may then infer with higher probability that ni is similarly a hyponym of nk ?despite never
having encountered the pair (ni , nk ) within a single sentence.
6.1 Coordinate Term Classification
Prior work for identifying coordinate terms includes automatic word sense clustering methods based on distributional similarity (e.g., [12, 14]) or on pattern-based techniques, specifically using the coordination pattern ?X, Y, and Z? (e.g., [2]). We construct both types of
classifier. First we construct a vector-space model similar to [12] using single MINIPAR
dependency links as our distributional features.6 We use the normalized similarity score
from this model for coordinate term classification. We evaluate this classifier on our handlabeled test set, where of 5,387 total pairs, 131 are labeled as ?coordinate?. For purposes
of comparison we construct a series of classifiers from WordNet, which make the binary
decision of determining whether two nouns are coordinate according to whether they share
a common ancestor within k nouns higher up in the hypernym taxonomy, for all k from 1
to 6. Also, we compare a simple pattern-based classifier based on the conjunction pattern,
which thresholds simply on the number of conjunction patterns found between a given pair.
Results of this experiment are shown in Table 4 and Figure 4.
The strong performance of the simple conjunction pattern model suggests that it may be
worth pursuing an extended pattern-based coordinate classifier along the lines of our hypernym classifier; for now, we proceed with our distributional similarity vector space model
(with a 16% relative F-score improvement over the conjunction model) in the construction
of a combined hypernym-coordinate hybrid classifier.
6.2 Hybrid hypernym-coordinate classification
We now combine our hypernym and coordinate models in order to improve hypernym classification. We define two probabilities of pair relationships between nouns: P (n i < nj ),
H
6
We use the same 6 million MINIPAR-parsed sentences used in our hypernym training set. Our
feature lexicon consists of the 30,000 most frequent noun-connected dependency edges. We construct
feature count vectors for each of the most frequently occurring 163,198 individual nouns. As in [12]
we normalize these feature counts with pointwise mutual information, and compute as our measure
of similarity the cosine coefficient between these normalized vectors.
Interannotator Agreement:
TREC+Wikipedia Hypernym-only Classifier (Logistic Regression):
TREC Hybrid Linear Interpolation Hypernym/Coordinate Model:
TREC Hypernym-only Classifier (Logistic Regression):
Best WordNet Classifier:
Hearst Patterns Classifier:
?And/Or Other? Pattern Classifier:
0.8318
0.3592
0.3268
0.2714
0.2339
0.1417
0.1386
Table 5: Maximum F-Score of hypernym classifiers on hand-labeled test set
representing the probability that noun ni has nj as an ancestor in its hypernym hierarchy,
and P (ni ? nj ), the probability that nouns ni and nj are coordinate terms, i.e., that they
C
share a common hypernym ancestor at some level. Defining the probability produced by
our best hypernym-only classifier as Pold (ni < nj ), and a probability obtained by normalH
izing the similarity score from our coordinate classifier as P (ni ? nj ), we apply a simple
C
linear interpolation scheme to compute a new hypernymy probability. Specifically, for each
pair of nouns (ni , nk ), we recompute the probability that nk is a hypernym of ni as:7
P
Pnew (ni < nk ) ? ?1 Pold (ni < nk ) + ?2 j P (ni ? nj )Pold (nj < nk )
H
7
H
C
H
Results
Our hand-labeled dataset allows us to compare our classifiers with WordNet and the previous feature-based methods, now using the human labels as ground truth. Figure 5 shows
the performance of each method in a precision/recall plot. We evaluated several classifiers
based on the WordNet hypernym taxonomy.8 The best WordNet-based results are plotted
in Figure 5. Our logistic regression hypernym-only model trained on the newswire corpora has a 16% relative F-score improvement over the best WordNet classifier, while the
combined hypernym/coordinate model has a 40% relative F-score improvement. Our bestperforming classifier is a hypernym-only model additionally trained on the Wikipedia corpus, with an expanded feature lexicon of 200,000 dependency paths; this classifier shows a
54% improvement over WordNet. In Table 5 we list the maximum F-scores of each method.
In Table 6 we analyze the disagreements between the highest F-score WordNet classifier
and our combined hypernym/coordinate classifier.9
8
Conclusions
Our experiments demonstrate that automatic methods can be competitive with WordNet
for the identification of hypernym pairs in newswire corpora. In future work we will use
the presented method to automatically generate flexible, statistically-grounded hypernym
taxonomies directly from corpora. These taxonomies will be made publicly available to
complement existing semantic resources.
7
We constrain our parameters ?1 , ?2 such that ?1 +?2 = 1; we set these parameters using 10-fold
cross-validation on our hand-labeled test set. For our final evaluation we use ? 1 = 0.7.
8
We tried all combinations of the following parameters: the maximum number of senses of a hyponym for which to find hypernyms, the maximum distance between the hyponym and its hypernym
in the WordNet taxonomy, and whether or not to allow synonyms. The WordNet model achieving the
maximum F-score uses only the first sense of a hyponym and allows a maximum distance of 4 links
between a hyponym and hypernym.
9
There are 31 such disagreements, with WordNet agreeing with the human labels on 5 and our hybrid model agreeing on the other 26. We additionally inspect the types of noun pairs where our model
improves upon WordNet, and find that at least 30% of our model?s improvements are not restricted
to Named Entities; given that the distribution of Named Entities among the labeled hypernyms in our
test set is over 60%, this gives us hope that our classifier will perform well at the task of hypernym
induction even in more general, non-newswire domains.
Type of Noun Pair
NE: Person
NE: Place
NE: Company
NE: Other
Not Named Entity:
Count
7
7
2
1
9
Example Pair
?John F. Kennedy / president?, ?Marlin Fitzwater / spokesman?
?Diamond Bar / city?, ?France / place?
?American Can / company?, ?Simmons / company?
?Is Elvis Alive / book?
?earthquake / disaster?, ?soybean / crop?
Table 6: Analysis of improvements over WordNet
Acknowledgments
We thank Kayur Patel, Mona Diab, Allison Buckley, and Todd Huffman for useful discussions and assistance annotating data. R. Snow is supported by an NDSEG Fellowship sponsored by the DOD and AFOSR. This work is also supported by the ARDA
AQUAINT program, and by the Department of the Interior/DARPA under contract number
NBCHD030010.
References
[1] Caraballo, S.A. (2001) Automatic Acquisition of a Hypernym-Labeled Noun Hierarchy from
Text. Brown University Ph.D. Thesis.
[2] Cederberg, S. & Widdows, D. (2003) Using LSA and Noun Coordination Information to Improve
the Precision and Recall of Automatic Hyponymy Extraction. Proc. of CoNLL-2003, pp. 111?118.
[3] Ciaramita, M. & Johnson, M. (2003) Supersense Tagging of Unknown Nouns in WordNet. Proc.
of EMNLP-2003.
[4] Ciaramita, M., Hofmann, T., & Johnson, M. (2003) Hierarchical Semantic Classification: Word
Sense Disambiguation with World Knowledge. Proc. of IJCAI-2003.
[5] Fellbaum, C. (1998) WordNet: An Electronic Lexical Database. Cambridge, MA: MIT Press.
[6] Girju, R., Badulescu A., & Moldovan D. (2003) Learning Semantic Constraints for the Automatic
Discovery of Part-Whole Relations. Proc. of HLT-2003.
[7] Harman, D. (1992) The DARPA TIPSTER project. ACM SIGIR Forum 26(2), Fall, pp. 26?28.
[8] Hearst, M. (1992) Automatic Acquisition of Hyponyms from Large Text Corpora. Proc. of the
Fourteenth International Conference on Computational Linguistics, Nantes, France.
[9] Hearst, M. & Sch?utze, H. (1993) Customizing a lexicon to better suit a computational task. In
Proc. of the ACL SIGLEX Workshop on Acquisition of Lexical Knowledge from Text.
[10] Lin, D. (1998) Dependency-based Evaluation of MINIPAR. Workshop on the Evaluation of
Parsing Systems, Granada, Spain
[11] Lin, D. & Pantel P. (2001) Discovery of Inference Rules for Question Answering. Natural
Language Engineering, 7(4), pp. 343?360.
[12] Pantel, P. (2003) Clustering by Committee. Ph.D. Dissertation. Department of Computing Science, University of Alberta.
[13] Pantel, P. & Ravichandran, D. (2004) Automatically Labeling Semantic Classes. Proc. of
NAACL-2004.
[14] Pereira, F., Tishby, N., & Lee, L. (1993) Distributional Clustering of English Words. Proc. of
ACL-1993, pp. 183?190.
[15] Ravichandran, D. & Hovy, E. (2002) Learning Surface Text Patterns for a Question Answering
system. Proc. of ACL-2002.
[16] Rennie J., Shih, L., Teevan, J., & Karger, D. (2003) Tackling the Poor Assumptions of Naive
Bayes Text Classifiers. Proc. of ICLM-2003.
[17] Riloff, E. & Shepherd, J. (1997) A Corpus-Based Approach for Building Semantic Lexicons.
Proc of EMNLP-1997.
[18] Roark, B. & Charniak, E. (1998) Noun-phrase co-occurerence statistics for semi-automaticsemantic lexicon construction. Proc. of ACL-1998, 1110?1116.
[19] Tseng, H. (2003) Semantic classification of unknown words in Chinese. Proc. of ACL-2003.
[20] Turney, P.D., Littman, M.L., Bigham, J. & Shanyder, V. (2003) Combining independent modules to solve multiple-choice synonym and analogy problems. Proc. of RANLP-2003, pp. 482?489.
[21] Widdows, D. (2003) Unsupervised methods for developing taxonomies by combining syntactic
and statistical information. Proc. of HLT/NAACL 2003, pp. 276?283.
| 2659 |@word version:1 tedious:1 tried:1 pold:3 hyponym:21 dramatic:1 initial:1 contains:2 fragment:1 score:15 series:1 daniel:1 karger:1 charniak:1 past:2 existing:1 stemmed:1 conjunct:1 tackling:1 parsing:1 john:1 happen:1 shakespeare:6 entrance:1 interannotator:5 hofmann:1 remove:1 designed:1 plot:2 sponsored:1 v:1 discovering:2 selected:3 indicative:3 dissertation:1 record:1 provides:1 recompute:1 node:3 lexicon:9 five:3 along:1 constructed:1 prove:1 consists:3 combine:2 paragraph:1 introduce:1 tagging:1 expected:1 roughly:1 frequently:4 multi:1 hypernym:86 detects:1 relying:1 alberta:1 automatically:13 company:3 increasing:1 becomes:1 project:1 classifies:1 notation:1 unrelated:2 spain:1 kind:2 differing:1 finding:1 marlin:1 nj:13 quantitative:1 every:2 classifier:61 subtype:1 lsa:1 arguably:1 engineering:1 treat:1 todd:1 severely:1 despite:1 path:23 interpolation:2 might:1 acl:5 conversely:1 collect:2 suggests:2 co:1 jurafsky:2 limited:3 statistically:1 averaged:2 directed:1 unique:3 acknowledgment:1 earthquake:1 rion:2 implement:1 word:12 pre:3 regular:1 interior:1 selection:1 ravichandran:2 lexical:2 regardless:1 focused:1 sigir:1 identifying:3 insight:1 rule:1 coordinate:29 justification:1 president:1 simmons:1 resp:1 hierarchy:4 construction:5 parser:2 us:1 agreement:6 element:1 expensive:1 rec:1 native:1 database:2 labeled:20 observed:2 distributional:6 module:1 capture:3 news:1 connected:1 highest:1 intuition:2 subtask:1 littman:1 trained:3 upon:1 triangle:1 darpa:2 articulated:1 train:2 separated:1 distinct:1 describe:1 labeling:1 quite:1 stanford:9 solve:1 rennie:2 annotating:1 statistic:1 unseen:2 syntactic:15 transform:1 final:1 propose:1 frequent:1 combining:2 pantel:3 hypernymy:6 normalize:1 los:1 ijcai:1 extending:2 satellite:2 produce:1 object:1 help:1 andrew:1 strong:1 coverage:1 c:2 implemented:1 indicate:1 implies:1 snow:2 human:2 engineered:2 assign:1 generalization:1 wall:1 desc:1 considered:1 ground:1 scope:1 achieves:1 utze:1 purpose:3 proc:15 diab:1 label:7 coordination:2 create:2 successfully:1 city:1 hope:1 mit:1 conjunction:10 improvement:9 attains:1 sense:8 inference:1 typically:2 relation:17 ancestor:6 france:2 nantes:1 classification:11 flexible:1 among:1 noun:68 mutual:1 once:2 never:1 having:1 ng:1 construct:4 extraction:1 identical:1 represents:2 broad:1 capitalize:1 unsupervised:1 nearly:1 future:1 np:35 word1:2 word2:2 girju:1 randomly:1 preserve:1 individual:2 consisting:3 maintain:1 suit:1 invariably:1 interest:1 highly:1 evaluation:7 introduces:1 allison:1 sens:2 activated:1 chain:1 edge:2 partial:1 conj:5 tree:9 re:2 plotted:1 bestperforming:1 instance:1 formalism:1 earlier:1 dev:1 phrase:3 cost:1 entry:1 subset:1 usefulness:1 dod:1 successful:1 johnson:2 tishby:1 reported:1 dependency:31 combined:3 confident:1 person:1 international:1 contract:1 lee:1 connecting:1 thesis:1 ndseg:1 containing:1 emnlp:2 soybean:1 book:1 american:1 style:1 concordance:1 includes:1 coefficient:1 performed:1 try:1 jason:1 analyze:2 linked:1 red:1 competitive:1 relied:1 bayes:5 ni:17 accuracy:2 publicly:1 hovy:1 judgment:1 identify:4 spaced:1 identification:2 polysemous:1 produced:1 worth:1 researcher:1 kennedy:1 hlt:2 against:1 acquisition:3 pp:6 associated:1 dataset:1 popular:1 recall:11 knowledge:7 improves:1 fellbaum:1 higher:5 originally:1 supervised:1 improved:1 formulation:1 evaluated:2 hand:18 parse:3 web:1 defines:1 logistic:10 building:2 naacl:2 brown:2 normalized:2 tagged:1 assigned:3 harman:1 symmetric:1 semantic:16 assistance:1 speaker:1 cosine:1 generalized:1 demonstrate:1 interface:1 meaning:1 novel:1 wikipedia:4 common:4 multinomial:2 exponentially:1 million:2 extend:2 linking:1 discussed:3 occurred:1 significant:1 expressing:1 cambridge:1 automatic:11 similarly:1 newswire:5 language:3 had:2 access:1 similarity:7 surface:1 add:1 labelers:2 base:1 recent:2 certain:2 rep:8 binary:9 scoring:1 additional:1 employed:1 determine:1 shortest:4 paradigm:1 redundant:2 semi:1 multiple:2 infer:1 determination:1 cross:3 lin:2 basic:1 regression:10 crop:1 bigham:1 metric:1 represent:2 grounded:1 disaster:1 addition:1 huffman:1 fellowship:1 interval:1 source:4 sch:1 rest:2 exhibited:1 file:2 shepherd:1 thing:1 member:1 arda:1 mod:9 obj:2 call:2 presence:2 split:1 variety:1 identified:1 reduce:1 intensive:1 angeles:1 whether:5 expression:1 motivated:1 six:1 distributing:1 speech:2 proceed:1 buckley:1 useful:5 customizing:2 listed:2 ang:1 desk:1 ten:1 encyclopedia:1 ph:2 category:3 generate:1 exist:1 dotted:1 dummy:2 blue:1 key:2 four:4 pnew:1 threshold:3 shih:1 achieving:1 drawn:1 neither:1 thresholded:1 vast:1 fourteenth:1 taxonomic:1 named:3 place:2 pursuing:1 electronic:1 thesaurus:2 disambiguation:1 decision:3 roark:1 conll:1 modifier:1 furniture:1 fold:3 encountered:1 occur:5 precisely:1 handful:1 constrain:1 alive:1 constraint:1 software:1 extremely:1 expanded:1 department:5 structured:2 according:2 developing:1 combination:1 poor:1 describes:1 across:3 smaller:1 agreeing:2 teevan:1 restricted:1 bucket:4 resource:1 previously:1 count:4 committee:1 generalizes:1 available:1 moldovan:1 apply:1 observe:1 hierarchical:1 disagreement:2 original:2 clustering:3 include:3 linguistics:2 parsed:2 build:3 especially:2 chinese:1 forum:1 noticed:1 already:1 added:1 occurs:3 question:2 distance:2 link:9 thank:1 entity:3 distributive:1 street:1 majority:1 participate:1 evenly:1 collected:1 tseng:1 reason:1 induction:2 ciaramita:2 pointwise:1 relationship:5 ratio:1 setup:1 taxonomy:15 unknown:2 canine:1 perform:2 diamond:1 inspect:1 optional:1 defining:1 extended:2 trec:6 mona:1 hearst:15 verb:1 pred:1 complement:3 pair:53 dog:1 lexico:10 sentence:12 vbe:2 nbchd030010:1 able:3 bar:1 below:1 pattern:61 program:1 built:2 including:2 power:2 suitable:1 greatest:1 natural:3 hybrid:5 representing:2 scheme:1 improve:3 ne:4 created:1 extract:4 naive:5 text:7 prior:1 literature:1 discovery:5 determining:2 relative:4 afosr:1 expect:1 recreating:1 analogy:1 versus:1 annotator:1 validation:3 tipster:1 article:3 classifying:1 share:3 granada:1 preposition:1 summary:2 supported:2 english:2 side:1 allow:1 wide:1 fall:1 sparse:1 distributed:1 boundary:1 dimension:1 evaluating:1 world:1 rich:1 glossary:1 author:6 made:2 instructed:1 patel:1 corpus:18 tuples:2 comma:1 table:12 additionally:2 nature:1 ca:3 constructing:2 domain:2 synonym:2 whole:1 prep:3 crafted:2 dump:1 depicts:1 slow:1 formalization:3 precision:9 inferring:1 pereira:1 wish:1 answering:2 herrick:3 specific:3 repeatable:1 symbol:1 list:4 exists:1 izing:1 workshop:2 rel:1 adding:2 false:1 occurring:2 nk:8 simply:4 likely:1 rediscover:1 labor:1 ordered:3 contained:2 applies:1 corresponds:2 truth:1 extracted:3 ma:1 acm:1 goal:3 marked:3 consequently:1 room:1 replace:1 determined:2 specifically:2 wordnet:38 called:1 total:2 elvis:1 experimental:1 turney:1 mark:1 evaluate:3 |
1,827 | 266 | 290
Viola
Neurally Inspired Plasticity in Oculomotor
Processes
Paul A. Viola
Artificial Intelligence Laboratory
M"assachusetts Institute of Technology
Cambridge, MA 02139
ABSTRACT
We have constructed a two axis camera positioning system which
is roughly analogous to a single human eye. This Artificial-Eye (Aeye) combines the signals generated by two rate gyroscopes with
motion information extracted from visual analysis to stabilize its
camera. This stabilization process is similar to the vestibulo-ocular
response (VOR); like the VOR, A-eye learns a system model that
can be incrementally modified to adapt to changes in its structure,
performance and environment. A-eye is an example of a robust sensory system that performs computations that can be of significant
use to the designers of mobile robots.
1
Introduction
We have constructed an "artificial eye" (A-eye), an autonomous robot that incorporates a two axis camera positioning system (figure 1). Like a the human oculomotor
system, A-eye can estimate the rotation rate of its body with a gyroscope and estimate the rotation rate of its "eye" by measuring image slip acr~ its "retina". Using
the gyroscope to sense rotation, A-eye attempts to stabilize its camera by driving
the camera motors to counteract body motion. The conversion of gyro output to
motor command is dependent on the characteristics of the gyroscope, the structure
of camera lensing system and the response of the motors. A correctly functioning stabilization system must model the characteristics of each of these external
variables.
Neurally Inspired Plasticity in Oculomotor Processes
Figure 1: The construction of A-eye can be viewed in rough analogy to the human
oculomotor system. In place of an eye, A-eye has a camera on a two axis positioning platform. In place of the circular canals of the inner ear, A-eye has two rate
gyroscopes that measure rotation in perpendicular axes.
Since camera motion implies stabilization error, A-eye uses a visual estimate of
camera motion to incrementally update its system model. \Vhen the camera is correctly stabilized there is no statistically significant slip. \Vhenever a particular gyro
measurement is associated with a result camera motion, A-eye makes an incremental change to its response to that particular measurement to reduce that error in
the future.
A-eye was built for two reasons: to facilitate the operation of complex visually
guided mobile robots and to explore the applicability of simple learning techniques
to the construction of a robust robot.
2
Autonomous Robots
An autonomous robot must function correctly for long periods of time without human intervention. It is certainly difficult to create an autonomous robot or process
that will function accurately, both initially and perpetually. To achieve such a goal,
autonomous processes must be able to adapt both to unforeseen aspects of the environment and inaccuracies in construction. One approach to attaining successful
autonomous performance would entail the full characterization of the robot's structure, its performance requirements, and its relationship with the environment. Since
clearly both the robot and its environment are susceptible to change any characterization could not be static. In contrast, our approach only partially categorizes
the robot's structure, environment, and task. Without more detailed information
initial performance is inaccurate. However, by using a measure of error in performance initially partial categorization can incrementally improved. In addition, a
change to system performance can be compensated continually. In this way the
extensive analysis and engineering that would be required to characterize, foresee
291
292
Viola
and circumvent variability can be greatly reduced.
3
The VOR
The oculomotor processes found in vertebrates are well studied. examples of adaptive, visually guided processing [Gou85]. The three oculomotor processes found
almost universally in vertebrates (the vestibulo-ocular response, the optokinetic
system, and the saccadic system), accurately perform ocular positioning tasks with
little or no conscious direction. The response times of these systems demonstrate
that little high level, "conscious", processing could take place. In a limited sense
these processes are autonomous, and it should come as no surprise that they are
quite plastic. Such plasticity is necessary to counteract the foreseeable changes in
the eye due to growth and aging and the unforeseeable changes due to illness and
injury.
The VOR works to counteract the motion of a creature in its environment. A correctly functioning VOR ensures that a creature "sees" as little unintended motion
as possible. Miles [FAM81] and others have demonstrated that the VOR is an
adaptive motor response, capable of significant recalibration in a matter of days.
Adaptation can be demonstrated by the use of inverting or magnifying spectacles.
While wearing these glasses the correct orbital motion of the eye, given a particular head motion, is significantly different from the normal response. Initially, the
response to head motion is an incorrect eye motion. With time eye motion begins
to approach the correct counteracting motion. This kind of adaptation allows an
animal to continue functioning in spite of injury or illness.
4
The Device
A-eye is a small autonomous robot that incorporates a CCD camera, a three wheel
base, a two axis pitch/yaw camera positioning platform, and two rate gyroscopes.
On board processing includes a Motorola microcontroller and 68020 based video
processing board. Including batteries, A-eye is a foot high cylinder that is 12 inches
wide. In its present configuration A-eye can run autonomously for up to three hours
(figure 2).
A-eye's goal is to learn how to keep its camera stable as its base trundles down
corridors. There are two sources of information regarding the motion of A-eye's
base: gyro rotation measurements and optical flow. Rate gyroscopes measure base
rotation rate directly. Visual analysis can be used to estimate motion by a number
of methods of varying complexity (see [Hil83] for a good overview). By attempting
to measure only camera rotation from slip complexity can be avoided. The simple
method we have chosen measures the slip of images across the retina.
4.1
Visual Rotation Estimation
Our approach to camera rotation estimation uses a pre-processing subunit commonly known as a "Reichard detector" which for clarity we will call a shift and
correlate .nit [PR73]. A shift and cOJTelate unit has as its inputs a set of samples
Neurally Inspired Plasticity in Oculomotor Processes
Figure 2: A photo of the current state of A-eye.
from a blurred area of the retina. It shifts these inputs spatially and correlates them
with a previous, unshifted set of inputs. \"hen two succeeding images are identical except for a spatial shift, the units which perform that shift respond strongly.
Clearly the activity of a shift and correlate anit contains information about retinal
motion. Due to the size and direction of shift, some detectors will be sensitive
to small motions, others large motions, and each will be sensitive to a particular
direction of motion.
The input from the shift and correlate units is used to build value-unit encoded
retinal velocity map, in which each unit is sensitive to a different direction and
range of velocities. The map has 9 units in a 3 by 3 grid (fig 3). To create such
a map, each of the shift and correlate uniu is connected to every map unit. By
moving the camera, displaced images that are examples of motion, are generated.
The motor command that generated this motion example corresponds to a unit
in the visual velocity map. Connection weights are updated by a standard least
squares learning rule. In operation, the most ac.tive unit represents the estimate of
visual motion.
4.2
Gyroscope Rotation Estimation
Contrary to first intuition, vertebrates do not rely on visual information to stabilize their eyes. Instead head rotation information measured by the inner ear, or
the vestibula, is used keep the eyes stable. Animals do not measure ocular motion
directly from visual information for two reasons: a) the response rates of photoreceptors prevent useful visual processing during rapid eye movements [Gou85] b) the
293
294
Viola
(s)CD0
808
0CDG)
Figure 3: The 9 unit velocity map has 1 unit for each of the 8 "chess moves" .
Base
Rotation
Rate
--.
Gyroscope
Transfa
Function
Inverting
Transfer
Function
Motor
Transfer
Function
......
Eye
Rotation
Rate
Figure 4: Open-loop control of ocular position based on gyroscope output.
required visual analysis takes approximately lOOms l . These difficulties combine to
prevent rapid response to unexpected head and body motions. A-eye is beset with
similar limitations and we have chosen a similar solution.
The output of the gyroscope is some function of head rotation rate. Stabilization
is achieved by driving the ocular motors directly in opposition to the measured
velocity (fig 4). This counteract rotation of the base in one direction by moving
the camera in the opposite direction. Such an open-loop system is very simple and
can perform well; they are unfortunately very reliant on proper calibration and
recalibration to maintain performance [Oga70).
A-eye maintains calibration information in the form of a function from gyroscope
output to motor velocity command. This function is an 8 unit gaussian radial basis
approximation network (TP89). Basis function approximation has excellent computational properties while representing wide variety of smooth functions. Weights
are modified with a simple least squares update rule, based errors in camera motion
detected visually.
5
Training A-eye
A-eye learns to perform the VOR in a two phase process. First, the measurement of
visual motion is calibrated to the generation of camera motion commands. Second,
lOcular following, the tendency to follow the motion of a !Cene in the &beence of head motion,
has a typical latency of lOOms [FM87).
Neurally Inspired Plasticity in Oculomotor Processes
the stabilizing motor responses to gyroscope measurements are approximated. This
approximation is modified based on a visual estimate of camera motion .
By observing motor commands and comparing them to the resulting visual motion,
a map from visual motion to appropriate motor command can be learned. To train
the visual motion map, A-eye performs a set of characteristic motions and observes
the results. Each motor command is categorized as one of the 9 distinct motions
encoded by the visual motion map. With each motion, the connections from shift
and correlate units to the visual motion map are updated so that issuing a motor
command results in activity in the correct visual motion unit. Because no reference
is made to external variables, this measure of visual motion is completely relative
to the function of the camera motors. The visual motion map plays the role of ertor
signal for later learning.
By observing both the gyroscope output and the visual response from head motion,
A-eye learns the appropriate compensating eye motion for all head motions. Eye
compensation motions are the result of motor commands generated by the approximation network applied to the gyroscope output. Incorrect responses will cause
visual mot.ion. This motion, as measured by the visual motion detector, is the error
signal that drives the modification of the approximation network. This is the heart
of the adaptation in the VOR.
5.1
Results
While training the motion detector and approximation network there are 5 training
events per second (the visual analysis takes about 200 msec). Training the visual
motion detector can take up to 10 minutes (in a few environments the weights refuse
to settle on the correct values). While it is possible to hand wire a detector that
is 95% accurate, most learned detectors worked well, attaining 85% accuracy. In
both cases, the detectors have the desirable capability of rejecting object motion
whenever there is actual camera motion (this is due to the global nature of the
analysis).
The approximation network converges to a function that performs well in minutes
(figure 5) . Analysis of the images generated by the camera leads us to bound the
cumulative error in rotation over a 1 minute trial at 5 degrees (we believe this
approaches the accuracy limitations inherent in the gyroscope).
An approach to reducing this gyroscope error involves yet another oculomotor pr~
cess: optokinetic nystagmus (OKN). This is the tendency for an otherwise undirected eye to follow visual motion in the absence of vestibular cues. A-eye's visual
motion map is in motor coordinates. By directing the camera in the opposite direction from observed motion, residual errors in VOR can be reduced.
6
Application
We claim that the stabilization that results from a correctly calibrated VOR is
useful both for navigation and scene analysis. A stable inertial reference can act
to assist tactical navigation when traversing rough terrain. Large body attitude
295
296
Viola
4000.
2000.
-60.
-40.
20.
40.
60.
-2000.
-4000.
Figure 5: A correct transfer function (rough) and the learned (smoother) approximation.
changes, that can result from such travel, make it difficult to maintain a navigational
bearing. However, when there exists a relatively stable inertial reference fr :lme less
analysis need be performed to predict or sense changes in bearing by other means.
The VOR is especially applicable to legged vehicles, where the terrain and the form
of locomotion can cause constant rapid changes in attitude [Rai89] [Ang89]. The
task of adapting conventional vision systems to such vehicles is formidable. As
the rate of pitching increases. the quality of video images degrade, while the task
of finding a correspondence between successive images will increase in complexity.
With the addition of the visual stabilization that A-eye can provide, an otherwise
complex visual analysis task can be much simplified.
7
Conclusions
A-eye is in part a response to the observation that static calibration is a disastrous
weakness. Static calibration not only forces an engineer to expend additional effort
at design time. it requires constant performance monitoring and recalibration. By
creating a device that monitors its own performance and adapts to changes. significant work can be saved in design and at numerous times during the lifetime of the
device.
A-eye is also in part a confirmation that simple. tractable and reliable learning
. mechanisms are sufficient to perform useful motor learning.
Finally. A-eye is in part a demonstration that useful visual processing can be performed in real-time with an reasonable amount of computation. This processing
yields the additional side-benefit of simplifying the complex task of visual recognition.
Neurally Inspired Plasticity in Oculomotor Processes
Acknowledgements
This report describes research done at the Artificial Intelligence Laboratory of the
Massachusetts Institute of Technology. Support for this research was provided by
Hughes Artificial Intelligence Center contract #SI-804475-D, the Office of Naval
Research contract NOOOI4-86-K-0685, and the Defense Advanced Research Projects
Agency under Office of Naval Research contract ~OOOI4-85-K-0124.
References
[Ang89] Colin Angle. Genghis, a six legged autonomous walking robot. Masterts
thesis, MIT, 1989.
[FAM81] S. G. Lisberger F . A. Miles. Plasticity in the vestibulo-ocular reflex: A
new hypothesis. Ann. Rev. Neurosci., 4:273-299,1981.
[FM87]
K. Kawano F.A. Miles. Visual stabilization of the eyes. TINS, 4(10):153158, 1987. Reference on Opto-kinetic nystagmus latency.
[Gou85] Peter Gouras. Oculomotor system. In James Schwartz Eric Kandel, editor, Principles of Neuroscience, chapter 34. Elsevier Science Publishing,
1985.
[Hil83]
Ellen C. Hildreth. The Measurement of Visual Motion. The MIT Press,
1983. Good book on the extraction of motion from edges.
[Oga70]
Katsuhiko Ogata. Modem Control Engineering. Prentice-Hall, Englewood Cliffs, N.J., 1970. Steady State Frequency Response (page 372).
[PR73]
T. Poggio and W. Reichard. Considerations on models of movement detection. Kybernetic, 13:223-227, 1973.
[Rai89]
Marc H. Raibert. Trotting, pacing, and bounding by a quadruped robot.
Journal of Biomechanics, 1989.
[TP89]
Federico Girosi Tomaso Poggio. A theory of networks for approximation
and learning. AI Memo 1140, MIT, 1989.
297
| 266 |@word trial:1 open:2 simplifying:1 initial:1 configuration:1 contains:1 unintended:1 current:1 comparing:1 si:1 yet:1 issuing:1 must:3 vor:11 plasticity:7 girosi:1 motor:17 succeeding:1 update:2 intelligence:3 cue:1 device:3 foreseeable:1 characterization:2 optokinetic:2 successive:1 constructed:2 corridor:1 incorrect:2 raibert:1 combine:2 orbital:1 rapid:3 tomaso:1 roughly:1 compensating:1 inspired:5 little:3 motorola:1 actual:1 vertebrate:3 begin:1 provided:1 project:1 formidable:1 kind:1 finding:1 every:1 act:1 growth:1 schwartz:1 reichard:2 unit:14 control:2 intervention:1 continually:1 engineering:2 aging:1 cd0:1 cliff:1 approximately:1 studied:1 pitching:1 limited:1 perpendicular:1 statistically:1 range:1 camera:25 hughes:1 area:1 significantly:1 adapting:1 pre:1 radial:1 spite:1 wheel:1 prentice:1 conventional:1 map:12 demonstrated:2 compensated:1 center:1 nit:1 stabilizing:1 rule:2 ellen:1 autonomous:9 coordinate:1 analogous:1 updated:2 construction:3 play:1 us:2 slip:4 hypothesis:1 locomotion:1 velocity:6 approximated:1 recognition:1 walking:1 observed:1 vestibula:1 role:1 ensures:1 connected:1 autonomously:1 movement:2 observes:1 intuition:1 environment:7 agency:1 complexity:3 battery:1 legged:2 eric:1 basis:2 completely:1 chapter:1 train:1 attitude:2 distinct:1 quadruped:1 artificial:5 detected:1 quite:1 encoded:2 otherwise:2 federico:1 acr:1 adaptation:3 fr:1 loop:2 achieve:1 adapts:1 requirement:1 categorization:1 incremental:1 converges:1 object:1 ac:1 measured:3 involves:1 implies:1 come:1 direction:7 guided:2 foot:1 correct:5 saved:1 stabilization:7 human:4 settle:1 pacing:1 unshifted:1 hall:1 normal:1 visually:3 predict:1 claim:1 driving:2 estimation:3 travel:1 applicable:1 sensitive:3 create:2 rough:3 clearly:2 mit:3 gaussian:1 modified:3 mobile:2 command:9 varying:1 office:2 ax:1 naval:2 greatly:1 contrast:1 sense:3 glass:1 elsevier:1 dependent:1 spectacle:1 inaccurate:1 initially:3 animal:2 platform:2 spatial:1 opto:1 categorizes:1 extraction:1 identical:1 represents:1 yaw:1 future:1 others:2 report:1 inherent:1 few:1 retina:3 phase:1 maintain:2 attempt:1 cylinder:1 detection:1 englewood:1 circular:1 certainly:1 weakness:1 navigation:2 accurate:1 edge:1 capable:1 partial:1 necessary:1 poggio:2 traversing:1 injury:2 measuring:1 applicability:1 successful:1 characterize:1 mot:1 calibrated:2 contract:3 unforeseen:1 thesis:1 ear:2 external:2 creating:1 book:1 attaining:2 retinal:2 stabilize:3 includes:1 tactical:1 matter:1 blurred:1 later:1 performed:2 vehicle:2 microcontroller:1 observing:2 maintains:1 capability:1 nystagmus:2 square:2 accuracy:2 characteristic:3 yield:1 inch:1 plastic:1 accurately:2 rejecting:1 monitoring:1 drive:1 detector:8 whenever:1 recalibration:3 frequency:1 ocular:7 james:1 associated:1 static:3 massachusetts:1 noooi4:1 inertial:2 kawano:1 day:1 follow:2 response:15 improved:1 foresee:1 done:1 strongly:1 lifetime:1 hand:1 incrementally:3 hildreth:1 quality:1 believe:1 facilitate:1 functioning:3 spatially:1 laboratory:2 vhen:1 mile:3 during:2 steady:1 demonstrate:1 performs:3 motion:56 image:7 consideration:1 rotation:16 overview:1 illness:2 significant:4 measurement:6 cambridge:1 ai:1 grid:1 moving:2 robot:13 entail:1 stable:4 calibration:4 base:6 own:1 cene:1 continue:1 additional:2 period:1 colin:1 signal:3 smoother:1 neurally:5 full:1 desirable:1 smooth:1 positioning:5 adapt:2 long:1 biomechanics:1 pitch:1 vision:1 achieved:1 ion:1 addition:2 source:1 undirected:1 contrary:1 incorporates:2 flow:1 call:1 counteracting:1 beset:1 variety:1 opposite:2 inner:2 reduce:1 regarding:1 shift:10 six:1 defense:1 assist:1 effort:1 peter:1 cause:2 useful:4 latency:2 detailed:1 amount:1 conscious:2 reduced:2 stabilized:1 designer:1 canal:1 neuroscience:1 correctly:5 per:1 monitor:1 clarity:1 prevent:2 ce:1 counteract:4 run:1 angle:1 respond:1 place:3 almost:1 reasonable:1 opposition:1 bound:1 correspondence:1 activity:2 worked:1 scene:1 aspect:1 attempting:1 optical:1 relatively:1 across:1 describes:1 rev:1 modification:1 chess:1 pr:1 heart:1 mechanism:1 tractable:1 photo:1 operation:2 appropriate:2 publishing:1 ccd:1 build:1 especially:1 move:1 saccadic:1 degrade:1 reason:2 relationship:1 demonstration:1 difficult:2 susceptible:1 unfortunately:1 disastrous:1 memo:1 design:2 proper:1 perform:5 conversion:1 wire:1 displaced:1 observation:1 modem:1 compensation:1 viola:5 subunit:1 variability:1 head:8 directing:1 tive:1 inverting:2 required:2 extensive:1 connection:2 learned:3 hour:1 inaccuracy:1 vestibular:1 able:1 refuse:1 oculomotor:11 gyroscope:17 built:1 including:1 navigational:1 video:2 reliable:1 event:1 difficulty:1 rely:1 circumvent:1 force:1 residual:1 advanced:1 representing:1 loom:2 technology:2 eye:45 numerous:1 axis:4 acknowledgement:1 hen:1 relative:1 generation:1 limitation:2 analogy:1 degree:1 sufficient:1 vestibulo:3 principle:1 editor:1 side:1 institute:2 wide:2 magnifying:1 expend:1 benefit:1 cumulative:1 sensory:1 commonly:1 perpetually:1 adaptive:2 universally:1 avoided:1 made:1 simplified:1 correlate:6 keep:2 global:1 photoreceptors:1 terrain:2 learn:1 transfer:3 robust:2 nature:1 confirmation:1 bearing:2 excellent:1 complex:3 marc:1 neurosci:1 bounding:1 paul:1 categorized:1 body:4 fig:2 creature:2 board:2 position:1 msec:1 kandel:1 tin:1 learns:3 ogata:1 down:1 minute:3 lme:1 exists:1 surprise:1 gyro:3 explore:1 reliant:1 visual:33 unexpected:1 partially:1 reflex:1 corresponds:1 lisberger:1 extracted:1 ma:1 kinetic:1 viewed:1 goal:2 ann:1 absence:1 change:10 typical:1 except:1 reducing:1 engineer:1 tendency:2 support:1 wearing:1 |
1,828 | 2,660 | An Information Maximization Model of
Eye Movements
Laura Walker Renninger, James Coughlan, Preeti Verghese
Smith-Kettlewell Eye Research Institute
{laura, coughlan, preeti}@ski.org
Jitendra Malik
University of California, Berkeley
[email protected]
Abstract
We propose a sequential information maximization model as a
general strategy for programming eye movements. The model
reconstructs high-resolution visual information from a sequence of
fixations, taking into account the fall-off in resolution from the
fovea to the periphery. From this framework we get a simple rule
for predicting fixation sequences: after each fixation, fixate next at
the location that minimizes uncertainty (maximizes information)
about the stimulus. By comparing our model performance to human
eye movement data and to predictions from a saliency and random
model, we demonstrate that our model is best at predicting fixation
locations. Modeling additional biological constraints will improve
the prediction of fixation sequences. Our results suggest that
information maximization is a useful principle for programming
eye movements.
1
In trod u ction
Since the earliest recordings [1, 2], vision researchers have sought to understand the
non-random yet idiosyncratic behavior of volitional eye movements. To do so, we
must not only unravel the bottom-up visual processing involved in selecting a
fixation location, but we must also disentangle the effects of top-down cognitive
factors such as task and prior knowledge. Our ability to predict volitional eye
movements provides a clear measure of our understanding of biological vision.
One approach to predicting fixation locations is to propose that the eyes move to
points that are ?salient?. Salient regions can be found by looking for centersurround contrast in visual channels such as color, contrast and orientation, among
others [3, 4]. Saliency has been shown to correlate with human fixation locations
when observers ?look around? an image [5, 6] but it is not clear if saliency alone
can explain why some locations are chosen over others and in what order. Task as
well as scene or object knowledge will play a role in constraining the fixation
locations chosen [7]. Observations such as this led to the scanpath theory, which
proposed that eye movement sequences are tightly linked to both the encoding and
retrieval of specific object memories [8].
1.1
Our Approach
We propose that during natural, active vision, we center our fixation on the most
informative points in an image in order to reduce our overall uncertainty about what
we are looking at. This approach is intuitive and may be biologically plausible, as
outlined by Lee & Yu [9]. The most informative point will depend on both the
observer?s current knowledge of the stimulus and the task. The quality of the
information gathered with each fixation will depend greatly on human visual
resolution limits. This is the reason we must move our eyes in the first place, yet it
is often ignored. A sequence of eye movements may then be understood within a
framework of sequential information maximization.
2
Human eye movements
We investigated how observers examine a novel shape when they must rely heavily
on bottom-up stimulus information. Because eye movements will be affected by the
task of the observer, we constructed a learn-discriminate paradigm. Observers are
asked to carefully study a shape and then discriminate it from a highly similar one.
2.1
Stimuli and Design
We use novel silhouettes to reduce the influence of object familiarity on the pattern
of eye movements and to facilitate our computations of information in the model.
Each silhouette subtends 12.5? to ensure that its entire shape cannot be characterized
with a single fixation.
During the learning phase, subjects first fixated a marker and then pressed a button
to cue the appearance of the shape which appeared 10? to the left or right of fixation.
Subjects maintained fixation for 300ms, allowing for a peripheral preview of the
object. When the fixation marker disappeared, subjects were allowed to study the
object for 1.2 seconds while their eye movements were recorded. During the
discrimination phase, subjects were asked to select the shape they had just studied
from a highly similar shape pair (Figure 1). Performance was near 75% correct,
indicating that the task was challenging yet feasible. Subjects saw 140 shapes and
given auditory feedback.
release fixation,
view object freely
(1200ms)
maintain fixation
(300ms)
Which shape is a match?
fixate, initiate trial
Figure 1. Temporal layout of a trial during the learning phase (left). Discrimination of
learned shape from a highly similar one (right).
2.2
Apparatus
Right eye position was measured with an SRI Dual Purkinje Image eye tracker while
subjects viewed the stimulus binocularly. Head position was fixed with a bitebar. A
25 dot grid that covered the extent of the presentation field was used for calibration.
The points were measured one at a time with each dot being displayed for 500ms.
The stimuli were presented using the Psychtoolbox software [10].
3
Model
We wish to create a model that builds a representation of a shape silhouette given
imperfect visual information, and which updates its representation as new visual
information is acquired. The model will be defined statistically so as to explicitly
encode uncertainty about the current knowledge of the shape silhouette. We will use
this model to generate a simple rule for predicting fixation sequences: after each
fixation, fixate next at the location that will decrease the model?s uncertainty as
much as possible. Similar approaches have been described in an ideal observer
model for reading [11], an information maximization algorithm for tracking
contours in cluttered images [12] and predicting fixation locations during object
learning [13].
3.1
Representing information
The information in silhouettes clearly resides at its contour, which we represent
with a collection of points and associated tangent orientations. These points and
their associated orientations are called edgelets, denoted e1, e2, ... eN, where N is the
total number of edgelets along the boundary. Each edgelet ei is defined as a triple
ei=(xi, yi, zi) where (xi, yi) is the 2D location of the edgelet and zi is the orientation
of the tangent to the boundary contour at that point. zi can assume any of Q possible
values 1, 2, ?, Q, representing a discretization of Q possible orientations ranging
from 0 to ? , and we have chosen Q=8 in our experiments. The goal of the model is
to infer the most likely orientation values given the visual information provided by
one or more fixations.
3.2
Updating knowledge
The visual information is based on indirect measurements of the true edgelet values
e1, e2, ... eN. Although our model assumes complete knowledge of the number N and
locations (xi, yi) of the edgelets, it does not have direct access to the orientations zi.1
Orientation information is instead derived from measurements that summarize the
local frequency of occurrence of edgelet orientations, averaged locally over a coarse
scale (corresponding to the spatial scale at which resolution is limited by the human
visual system). These coarse measurements provide indirect information about
individual edgelet orientations, which may not uniquely determine the orientations.
We will use a simple statistical model to estimate the distribution of individual
orientation values conditioned on this information.
Our measurements are defined to model the resolution limitations of the human
visual system, with highest resolution at the fovea and lower resolution in the
1
Although the visual system does not have precise knowledge of location coordinates, the
model is greatly simplified by assuming this knowledge. It is reasonable to expect that
location uncertainty will be highly correlated with orientation uncertainty, so that the
inclusion of location should not greatly affect the model's decisions of where to fixate next.
periphery. Distance to the fovea is rmeasured as eccentricity E, the visual angle
between any point and the fovea. If x = ( x, y ) is the location of a point in an image
r
and f = ( f x , f y ) is the fixation (i.e. foveal) location in the image then the
r
eccentricity is E = xr ? f , measured in units of visual degrees. The effective
resolution of orientation discrimination falls with increasing eccentricity as
r (E ) = FPH ( E + E 2 ) where r(E) is an effective radius over which the visual system
spatially pools information and FPH =0.1 and E2=0.8 [14].
Our model represents pooled information as a histogram of edge orientations within
the effective radius. For each edgelet ei we define the histogram of all edgelet
orientations ej within radius ri = r(E) of ei , where E is the eccentricity of xri = ( xi , yi )
r
r
relative to the current fixation f , i.e. E = xri ? f . To define the histogram more
precisely we will introduce the neighborhood set Ni of all indices j corresponding to
r r
edgelets within radius ri of ei : N i = all j s.t. xi ? x j ? ri , with number of
{
}
neighborhood edgelets |Ni|. The (normalized) histogram centered at edgelet ei is then
defined as
hiz =
1
Ni
??
j?N i
z,z j
,
which is the proportion of edgelet orientations that assume value z in the
(eccentricity-dependent) neighborhood of edgelet ei.2
Figure 2. Relation between eccentricity E and radius r(E) of the neighborhood (disk)
which defines the local orientation histogram (hiz ). Left and right panels show two
fixations for the same object.
Up to this point we have restricted ourselves to the case of a single fixation. To
designate a sequence of multiple fixations we will index them byrk=1, 2, ?, K (for
K total fixations). The k th fixation location is denoted by f ( k ) = ( f xk , f yk ) . The
quantities ri , Ni and hiz depend on fixation location and so to make this dependence
explicit we will augment them with superscripts as ri(k ) , N i(k ) , and hiz(k ) .
2
? x, y is the Kronecker delta function, defined to equal 1 if x = y and 0 if x ? y .
Now we describe the statistical model of edgelet orientations given information
obtained from multiple fixations. Ideally we would like to model the exact
distribution
of
orientations
conditioned
on
the
histogram
data:
(1)
( 2)
(K )
(k )
represents
all
histogram
,
where
P(zi , z 2 , ... z N | {hiz }, {hiz },K, {hiz })
{hiz }
r
components z at every edgelet ei for fixation f (k ) . This exact distribution is
intractable, so we will use a simple approximation. We assume the distribution
factors over individual edgelets:
N
P(zi , z 2 , ... z N | {hiz(1) }, {hiz( 2 ) },K, {hiz( K ) }) = ? g i(zi )
i =1
where gi(zi) is the marginal distribution of orientation zi. Determining these
marginal distributions is still difficult even with the factorization assumption, so we
K
will make an additional approximation: g (z ) = 1
hiz( k ) , where Zi is a suitable
?
i i
Z i k =1
(k )
normalization factor. This approximation corresponds to treating hiz as a likelihood
function over z, with independent likelihoods for each fixation k. While the
approximation has some undesirable properties (such as making the marginal
distribution gi(zi) more peaked if the same fixation is made repeatedly), it provides a
simple mechanism for combining histogram evidence from multiple, distinct
fixations.
3.3
Selecting the next fixation
r ( K +1)
Given the past K fixations, the next fixation f
is chosen to minimize the model
r ( K +1)
entropy of the edgelet orientations. In other words, f
is chosen to minimize
r ( K +1)
H( f
) = entropy[ P(zi , z2 , ... z N | {hiz(1) }, {hiz( 2 ) },K, {hiz( K +1) })] , where the entropy of a
distribution P(x) is defined as ? ? P( x) log P ( x) . In practice, we minimize the
x
r
entropy by evaluating it across a set of candidate locations f ( K +1) which forms a
regularly sampled grid across the image.3 We note that this selection rule makes
decisions that depend, in general, on the full history of previous K fixations.
4
Results
Figure 3 shows an example of one observer?s eye movements superimposed over the
shape (top row), the prediction from a saliency model (middle row) [3] and the
prediction from the information maximization model (bottom row). The information
maximization model updates its prediction after each fixation.
An ideal sequence of fixations can be generated by both models. The saliency model
selects fixations in order of decreasing salience. The information maximization
model selects the maximally informative point after incorporating information from
the previous fixations. To provide an additional benchmark, we also implemented a
3
This rule evaluates the entropy resulting from every possible next fixation before making a
decision. Although this rule is suitable for our modeling purposes, it would be inefficient to
implement in a biological or machine vision system. A practical decision rule would use
current knowledge to estimate the expected (rather than actual) entropy.
Figure 3. Example eye movement pattern, superimposed over the stimulus (top row),
saliency map (middle row) and information maximization map (bottom row).
model that selects fixations at random. One way to quantify the performance is to
map a subject?s fixations onto the closest model predicted fixation locations,
ignoring the sequence in which they were made. In this analysis, both the saliency
and information maximization models are significantly better than random at
predicting candidate locations (p < 0.05; t-test) for three observers (Figure 4, left).
The information maximization model performs slightly but significantly better than
the saliency model for two observers (lm, kr). If we match fixation locations while
retaining the sequence, errors become quite large, indicating that the models cannot
account for the observed behavior (Figure 4, right).
Sequence Error
Visual Angle (deg)
Location Error
R S I
R S I
R S I
R S I
R S I
R S I
Figure 4. Prediction error of three models: random (R), saliency (S) and information
maximization (I) for three observers (pv, lm, kr). The left panel shows the error in
predicting fixation locations, ignoring sequence. The right panel shows the error when
sequence is retained before mapping. Error bars are 95% confidence intervals.
The information maximization model incorporates resolution limitations, but there
are further biological constraints that must be considered if we are to build a model
that can fully explain human eye movement patterns. First, saccade amplitudes are
typically around 2-4? and rarely exceed 15? [15]. When we move our eyes, the
image of the visual world is smeared across the retina and our perception of it is
actively suppressed [16]. Shorter saccade lengths may be a mechanism to reduce
this cost. This biological constraint would cause a fixation to fall short of the
prediction if it is distant from the current fixation (Figure 5).
Figure 5. Cost of moving the eyes. Successive fixations may fall short of the maximally
salient or informative point if it is very distant from the current fixation.
Second, the biological system may increase its sampling efficiency by planning a
series of saccades concurrently [17, 18]. Several fixations may therefore be made
before sampled information begins to influence target selection. The information
maximization model currently updates after each fixation. This would create a
discrepancy in the prediction of the eye movement sequence (Figure 6).
Figure 6. Three fixations are made to a location that is initially highly informative
according to the information maximization model. By the fourth fixation, the subject
finally moves to the next most informative point.
5
D i s c u s s i on
Our model and the saliency model are using the same image information to
determine fixation locations, thus it is not surprising that they are roughly similar in
their performance of predicting human fixation locations. The main difference is
how we decide to ?shift attention? or program the sequence of eye movements to
these locations. The saliency model uses a winner-take-all and inhibition-of-return
mechanism to shift among the salient regions. We take a completely different
approach by saying that observers adopt a strategy of sequential information
maximization. In effect, the history of where we have been matters because our
model is continually collecting information from the stimulus. We have an implicit
?inhibition-of-return? because there is little to be gained by revisiting a point.
Second, we attempt to take biological resolution limits into account when
determining the quality of information gained with each fixation. By including
additional biological constraints such as the cost of making large saccades and the
natural time course of information update, we may be able to improve our prediction
of eye movement sequences.
We have shown that the programming of eye movements can be understood within a
framework of sequential information maximization. This framework is portable to
any image or task. A remaining challenge is to understand how different tasks
constrain the representation of information and to what degree observers are able to
utilize the information.
Acknowledgments
Smith-Kettlewell Eye Research Institute, NIH Ruth L. Kirschstein NRSA, ONR #N0001401-1-0890, NSF #IIS0415310, NIDRR #H133G030080, NASA #NAG 9-1461.
References
[1] Buswell (1935). How people look at pictures. Chicago: The University of Chicago Press.
[2] Yarbus (1967). Eye movements and vision. New York: Plenum Press.
[3] Itti & Koch (2000). A saliency-based search mechanism for overt and covert shifts of
visual attention. Vision Research, 40, 1489-1506.
[4] Kadir & Brady (2001). Scale, saliency and image description. International Journal of
Computer Vision, 45(2), 83-105.
[5] Parkhurst, Law, and Niebur (2002). Modeling the role of salience in the allocation of
overt visual attention. Vision Research, 42(1), 107-123.
[6] Nothdurft (2002). Attention shifts to salient targets. Vision Research, 42, 1287-1306.
[7] Oliva, Torralba, Castelhano & Henderson (2003). Top-down control of visual attention in
object detection. Proceedings of the IEEE International Conference on Image Processing,
Barcelona, Spain.
[8] Noton & Stark (1971). Scanpaths in eye movements during pattern perception. Science,
171, 308-311.
[9] Lee & Yu (2000). An information-theoretic framework for understanding saccadic
behaviors. Advanced in Neural Processing Systems, 12, 834-840.
[10] Brainard (1997). The psychophysics toolbox. Spatial Vision, 10 (4), 433-436.
[11] Legge, Hooven, Klitz, Mansfield & Tjan (2002). Mr.Chips 2002: new insights from an
ideal-observer model of reading. Vision Research, 42, 2219-2234.
[12] Geman & Jedynak (1996). An active testing model for tracking roads in satellite images. IEEE
Trans. Pattern Analysis and Machine Intel, 18(1), 1-14.
[13] Renninger & Malik (2004). Sequential information maximization can explain eye
movements in an object learning task. Journal of Vision, 4(8), 744a.
[14] Levi, Klein & Aitesbaomo (1985). Vernier acuity, crowding and cortical magnification.
Vision Research, 25(7), 963-977.
[15] Bahill, Adler & Stark (1975). Most naturally occurring human saccades have
magnitudes of 15 degrees or less. Investigative Ophthalmology, 14, 468-469.
[16] Burr, Morrone & Ross (1994). Selective suppression of the magnocellular visual
pathway during saccadic eye movements. Nature, 371, 511-513.
[17] Caspi, Beutter & Eckstein (2004). The time course of visual information accrual guiding
eye movement decisions. Proceedings of the Nat?l Academy of Science, 101(35), 13086-90.
[18] McPeek, Skavenski & Nakayama (2000). Concurrent processing of saccades in visual
search. Vision Research, 40, 2499-2516.
| 2660 |@word trial:2 middle:2 sri:1 proportion:1 disk:1 pressed:1 crowding:1 foveal:1 series:1 selecting:2 past:1 current:6 comparing:1 discretization:1 z2:1 surprising:1 yet:3 must:5 chicago:2 distant:2 informative:6 shape:12 treating:1 update:4 discrimination:3 alone:1 cue:1 xk:1 coughlan:2 smith:2 short:2 provides:2 coarse:2 location:28 successive:1 org:1 yarbus:1 along:1 constructed:1 direct:1 become:1 kettlewell:2 fixation:59 pathway:1 burr:1 introduce:1 acquired:1 expected:1 roughly:1 behavior:3 examine:1 planning:1 decreasing:1 actual:1 little:1 increasing:1 provided:1 begin:1 spain:1 maximizes:1 panel:3 what:3 minimizes:1 brady:1 temporal:1 berkeley:2 every:2 collecting:1 control:1 unit:1 continually:1 before:3 understood:2 local:2 apparatus:1 limit:2 encoding:1 studied:1 challenging:1 limited:1 factorization:1 statistically:1 averaged:1 jedynak:1 practical:1 acknowledgment:1 testing:1 practice:1 implement:1 xr:1 significantly:2 word:1 confidence:1 road:1 suggest:1 get:1 cannot:2 undesirable:1 selection:2 onto:1 preeti:2 influence:2 map:3 center:1 layout:1 attention:5 cluttered:1 unravel:1 renninger:2 resolution:10 rule:6 insight:1 coordinate:1 plenum:1 target:2 play:1 heavily:1 exact:2 programming:3 us:1 magnification:1 updating:1 geman:1 bottom:4 role:2 observed:1 revisiting:1 region:2 movement:24 decrease:1 highest:1 yk:1 legge:1 mcpeek:1 asked:2 ideally:1 depend:4 efficiency:1 completely:1 indirect:2 chip:1 distinct:1 investigative:1 effective:3 describe:1 ction:1 neighborhood:4 quite:1 plausible:1 kadir:1 ability:1 gi:2 superscript:1 sequence:16 propose:3 combining:1 academy:1 intuitive:1 description:1 eccentricity:6 satellite:1 disappeared:1 object:10 brainard:1 measured:3 implemented:1 predicted:1 quantify:1 radius:5 correct:1 centered:1 human:9 biological:8 designate:1 around:2 tracker:1 considered:1 koch:1 mapping:1 predict:1 lm:2 sought:1 adopt:1 torralba:1 purpose:1 overt:2 currently:1 ross:1 saw:1 concurrent:1 create:2 smeared:1 clearly:1 concurrently:1 rather:1 ej:1 earliest:1 encode:1 release:1 derived:1 acuity:1 verghese:1 likelihood:2 superimposed:2 greatly:3 contrast:2 suppression:1 dependent:1 entire:1 typically:1 initially:1 relation:1 selective:1 selects:3 overall:1 among:2 orientation:22 dual:1 psychtoolbox:1 denoted:2 augment:1 retaining:1 spatial:2 psychophysics:1 marginal:3 field:1 equal:1 sampling:1 represents:2 look:2 yu:2 peaked:1 discrepancy:1 others:2 stimulus:8 retina:1 tightly:1 individual:3 phase:3 ourselves:1 maintain:1 preview:1 attempt:1 detection:1 tjan:1 highly:5 henderson:1 edge:1 trod:1 shorter:1 modeling:3 purkinje:1 maximization:18 cost:3 eec:1 adler:1 international:2 lee:2 off:1 pool:1 recorded:1 reconstructs:1 castelhano:1 cognitive:1 laura:2 inefficient:1 itti:1 return:2 stark:2 actively:1 account:3 pooled:1 parkhurst:1 matter:1 jitendra:1 explicitly:1 view:1 observer:13 linked:1 minimize:3 fph:2 ni:4 gathered:1 saliency:13 niebur:1 researcher:1 history:2 explain:3 evaluates:1 frequency:1 involved:1 james:1 fixate:4 e2:3 associated:2 naturally:1 sampled:2 auditory:1 knowledge:9 color:1 amplitude:1 carefully:1 nasa:1 maximally:2 just:1 implicit:1 ei:8 marker:2 defines:1 quality:2 facilitate:1 effect:2 normalized:1 true:1 spatially:1 during:7 uniquely:1 maintained:1 m:4 complete:1 demonstrate:1 theoretic:1 performs:1 covert:1 ranging:1 image:13 novel:2 nih:1 winner:1 measurement:4 outlined:1 grid:2 inclusion:1 had:1 dot:2 moving:1 calibration:1 access:1 inhibition:2 disentangle:1 closest:1 periphery:2 onr:1 buswell:1 yi:4 additional:4 mr:1 freely:1 determine:2 paradigm:1 multiple:3 full:1 infer:1 match:2 characterized:1 retrieval:1 e1:2 prediction:9 mansfield:1 oliva:1 vision:14 histogram:8 represent:1 normalization:1 interval:1 walker:1 scanpaths:1 recording:1 subject:8 regularly:1 incorporates:1 near:1 edgelet:13 ideal:3 constraining:1 exceed:1 edgelets:6 affect:1 zi:12 reduce:3 imperfect:1 shift:4 york:1 cause:1 repeatedly:1 scanpath:1 ignored:1 useful:1 clear:2 covered:1 locally:1 generate:1 nsf:1 delta:1 klein:1 affected:1 levi:1 salient:5 nothdurft:1 utilize:1 volitional:2 button:1 angle:2 uncertainty:6 fourth:1 place:1 saying:1 reasonable:1 decide:1 decision:5 constraint:4 precisely:1 kronecker:1 constrain:1 scene:1 software:1 ri:5 according:1 peripheral:1 across:3 slightly:1 ophthalmology:1 suppressed:1 biologically:1 making:3 restricted:1 binocularly:1 mechanism:4 initiate:1 occurrence:1 top:4 assumes:1 ensure:1 remaining:1 hiz:16 build:2 magnocellular:1 malik:3 move:4 quantity:1 strategy:2 saccadic:2 dependence:1 fovea:4 distance:1 centersurround:1 extent:1 portable:1 reason:1 assuming:1 length:1 ruth:1 index:2 retained:1 difficult:1 idiosyncratic:1 vernier:1 xri:2 design:1 ski:1 allowing:1 observation:1 benchmark:1 displayed:1 nrsa:1 looking:2 head:1 precise:1 pair:1 eckstein:1 toolbox:1 california:1 learned:1 barcelona:1 trans:1 able:2 bar:1 pattern:5 perception:2 appeared:1 reading:2 summarize:1 challenge:1 program:1 including:1 memory:1 suitable:2 natural:2 rely:1 predicting:8 advanced:1 representing:2 improve:2 eye:32 picture:1 prior:1 understanding:2 tangent:2 determining:2 relative:1 law:1 fully:1 expect:1 limitation:2 allocation:1 triple:1 degree:3 principle:1 row:6 course:2 salience:2 understand:2 institute:2 fall:4 taking:1 feedback:1 boundary:2 cortical:1 evaluating:1 world:1 contour:3 resides:1 collection:1 made:4 simplified:1 correlate:1 silhouette:5 deg:1 active:2 fixated:1 nag:1 xi:5 morrone:1 search:2 why:1 channel:1 learn:1 nature:1 ignoring:2 nakayama:1 investigated:1 main:1 allowed:1 intel:1 en:2 position:2 pv:1 wish:1 explicit:1 guiding:1 candidate:2 down:2 familiarity:1 specific:1 evidence:1 intractable:1 incorporating:1 sequential:5 kr:2 gained:2 magnitude:1 nat:1 conditioned:2 occurring:1 entropy:6 led:1 appearance:1 likely:1 visual:22 tracking:2 saccade:6 corresponds:1 caspi:1 viewed:1 presentation:1 goal:1 feasible:1 called:1 total:2 discriminate:2 indicating:2 select:1 rarely:1 people:1 correlated:1 |
1,829 | 2,661 | Expectation Consistent Free Energies for
Approximate Inference
Manfred Opper
ISIS
School of Electronics and
Computer Science
University of Southampton
SO17 1BJ, United Kingdom
[email protected]
Ole Winther
Informatics and
Mathematical Modelling
Technical University of Denmark
DK-2800 Lyngby, Denmark
[email protected]
Abstract
We propose a novel a framework for deriving approximations for intractable probabilistic models. This framework is based on a free energy
(negative log marginal likelihood) and can be seen as a generalization
of adaptive TAP [1, 2, 3] and expectation propagation (EP) [4, 5]. The
free energy is constructed from two approximating distributions which
encode different aspects of the intractable model such a single node constraints and couplings and are by construction consistent on a chosen set
of moments. We test the framework on a difficult benchmark problem
with binary variables on fully connected graphs and 2D grid graphs. We
find good performance using sets of moments which either specify factorized nodes or a spanning tree on the nodes (structured approximation).
Surprisingly, the Bethe approximation gives very inferior results even on
grids.
1
Introduction
The development of tractable approximations for the statistical inference with probabilistic data models is of central importance in order to develop their full potential. The most
prominent and widely developed [6] approximation technique is the so called Variational
Approximation (VA) in which the true intractable probability distribution is approximated
by the closest one in a tractable family. The most important tractable families of distributions are multivariate Gaussians and distributions which factorize in all or in certain groups
of variables [7]. Both choices have their drawbacks. While factorizing distributions neglect
correlations, multivariate Gaussians allow to retain a significant amount of dependencies
but are restricted to continuous random variables which have the entire real space as their
natural domain (otherwise KL divergences becomes infinite).
More recently a variety of non variational approximations have been developed which can
be understood from the idea of global consistency between local approximations. E.g.,
in the Bethe?Kikuchi approach [8] the local neighborhood of each variable in a graphical
model is implicitly approximated by a tree-like structure. Consistency is achieved by the
matching of marginal distributions at the connecting edges of the graph. Thomas Minka?s
Expectation Propagation (EP) framework seems to provide a general framework for developing and unifying such consistency approximations [4, 5]. Although the new frameworks
have led to a variety of promising applications, often outperforming VA schemes, the unsatisfactory division between the treatment of constrained and unconstrained, continuous
random variables seems to persist.
In this paper we propose an alternative approach which we call the expectation consistent
(EC) approximation which is not plagued by this problem. We require consistency between
two complimentary global approximations (say, a factorizing & a Gaussian one) to the same
probabilistic model which may have different support. Our method is a generalization of
the adaptive TAP approach (ADATAP) [2, 3] developed for inference on densely connected
graphical models which has been applied successfully to a variety of problems ranging from
probabilistic ICA over Gaussian process models to bootstrap methods for kernel machines.
2
Approximative inference
We consider the problem of computing expectations, i.e. certain sums or integrals involving
a probability distribution with density
p(x) =
1
f (x) ,
Z
(1)
for
R a vector of random variables x = (x1 , x2 , . . . , xN ) with the partition function Z =
dxf (x). We assume that the necessary exact operations are intractable, where the intractability arises either because the necessary sums are over a too large number of variables or because multivariate integrals cannot be evaluated exactly. In a typical scenario,
f (x) is expressed as a product of two functions
f (x) = f1 (x)f2 (x)
(2)
with f1,2 (x) ? 0, where f1 is ?simple? enough to allow for tractable computations.
The idea of many approximate inference methods is to approximate the ?complicated?
part f2 (x) by replacing
it with a ?simpler?
function, say of some exponential form
P
K
T
exp ? g(x) ? exp
j=1 ?j gj (x) . The vector of functions g is chosen in such a
way that the desired sums or integrals can be calculated in an efficient way and the parameters ? are adjusted to optimize certain criteria. Hence, the word tractability should always
be understood as relative to some approximating set of functions g.
Our novel framework of approximation will be restricted to problems, where both parts f 1
and f2 can be considered as tractable relative to some suitable g, and the intractability of
the density p arises from forming their product. Take, as an example, the density (with
respect to the Lebesgue measure in RN ) given by
?
?
Y
X
p(x) =
?? (x? ) exp ?
xi Jij xj ? ,
(3)
?
i<j
where the x? denote tractable potentials defined on disjoint subsets of variables x? .
In order to have a non-trivial problem, the ?? should be a non-Gaussian function.
One may approximate
p(x) by a factorizing distribution, thereby replacing f 2 (x) ?
P
exp
i<j xi Jij xj by some function which factorizes in the components x i . AlternaQ
tively, one can consider replacing f1 (x) = i ?i (xi ) by a Gaussian function to make
the whole distribution Gaussian. Both approximations are not ideal. The first completely
neglects correlations of the variables but leads to marginal distributions of the x i , which
might qualitatively resemble the non Gaussian shape of the true marginal. The second
one neglects the non Gaussian effects but incorporates correlations which might be used
in order to approximate the two variable covariance functions. While within the VA both
approximations would appear independent from each others, we will, in the following develop an approach for combining two approximations which ?communicate? by matching
the corresponding expectations of the functions g(x). We do not have to assume that either
choice is a reasonably good approximation for the global joint density p(x) as done in the
VA. In fact, we apply the approach to a case where the KL divergence between one of them
and p is even infinite!
3
Gibbs free energy
Free energies (FE) provide a convenient formalism for dealing with probabilistic approximation problems. In this framework, the true, intractable distribution p(x) = f (x)
Z is
implicitly characterized as the solution of an optimization problem defined through the the
relative entropy or KL divergence
Z
q(x)
KL(q, p) = dx q(x) ln
(4)
p(x)
between p and other trial distributions q. In contrast to the usual formulation of the VA,
where one minimizes the KL divergence directly within a tractable family, it is more convenient to consider the following two stage optimization process. In the first step, one
constrains the trial distributions q by fixing the values of a set of generalized moments
hg(x)iq . This will be helpful later to enable the communication between approximations.
?) as
We define the Gibbs Free Energy G(?
?) = min {KL(q, p) | hg(x)iq = ? } ? ln Z .
G(?
q
(5)
We have subtracted the term ln Z to make the expression independent of the intractable
partition function Z. In a second step, the moments of the distribution and also the partition
function Z are found within the same approach by relaxing the constraints and further
?) with respect to the ?.
minimizing G(?
?) = ? ln Z
min G(?
?
and
?) .
hgi = argmin G(?
(6)
?
We will next give a short summary of properties of the Gibbs free energy (GFE). The
optimizing density in (5) is given by
f (x)
q(x) =
exp ?T g(x) ,
(7)
Z(?)
?) is
with a normalizing partition function Z(?). The set of Lagrange parameters ? = ?(?
chosen such that the conditions hg(x)iq = ? are fulfilled, i.e. ? satisfies
? ln Z(?)
=?.
??
(8)
Inserting the optimizing distribution eq. (7) into the definition of the Gibbs free energy eq.
(5), we get the explicit expression:
n
o
?) = ? ln Z(?(?
?)) + ?T (?
?)?
? = max ? ln Z(?) + ?T ? ,
G(?
(9)
?
i.e., G is the Legendre transform or dual of ? ln Z(?). Hence, G is a convex function of
?)
its arguments and ?G(?
? = ?.
??
?) can be used to generate moments, e.g.
G(?
?)
? 2 G(?
???
?T
??
=
??
=
?T
??
?T
??
??
?1
?1
= hg(x)gT (x)i ? hg(x)ihg(x)iT
, (10)
where the expectations are over the density eq. (7).
The derivative with respect to a parameter t contained in the probability density p(x|t) =
f (x,t)
Zt can be calculated using (9) and (8) as
?, t)
dG(?
? ln Z(?, t)
? ln Z(?, t) d?T
? ln Z(?, t)
=?
+ ??
=?
.
dt
?t
??
dt
?t
(11)
The important message is that we only need to take the explicit t dependence into account,
i.e. we can keep ? fixed upon differentiation.
3.1
Free energy examples
Here we give the free energies for three tractable models and choices of moments that will
be used subsequently in the free energy framework.
Q
Completely factorized, i.e. p(x) = i ?i (xi ). For simplicity we will consider biased
binary variables: ?i (xi ) = [?(xi + 1) + ?(xi ? 1)]e?i xi and fix the first moments m = hxi.
Denoting the conjugate Lagrange parameters by ?:
X
G(m) =
Gi (mi )
with Gi (mi ) = max {? ln Zi (?i ) + mi ?i }
(12)
?i
i
with Zi (?i ) =
R
dxi ?i (?)e?i xi = 2 cosh(?i + ?i ).
Tree-connected graph. For the case where either the couplings and the moments together define a tree-connected graph, we can write the free energy in term of single- and
two-node free energies. Considering again completely factorized binary variables, all nontrivial moments on the graph (ij) ? G are the means m and correlations of linked nodes
Mij = hxi xj i:
X
X
(1 ? ni )Gi (mi ) ,
(13)
G(m, {Mij }(ij)?G ) =
Gij (mi , mj , Mij ) +
(ij)?G
i
where Gij (mi , mj , Mij ) is the two-node free energy defined in a similar fashion as the
one-node free energy, ni the number of links to node i and Gi (mi ) is the one-node free
energy.
Gaussian distribution. We set ? = (m, M) with all first moments m and an arbitrary
subset of second moments M for a Gaussian model ?i (xi ) ? exp[ai xi ? b2i x2i ] and p(x)
given by eq. (3). We introduce conjugate variables ? and ??/2. ? can be eliminated
analytically, whereas we get a log-determinant maximization problem for ?:
G(m, M)
1
1X
= ? mT Jm ? mT a +
Mii bi
2
2 i
1
1
T
ln det(? ? J) ? Tr ?(M ? mm ) .
+ max
?
2
2
(14)
4
Exact interpolation representation
If the density p factors into a tractable f1 and an intractable part f2 , according to eq. (2),
we can construct a representation of the Gibbs free energy which also separates into two
corresponding parts. We treat f2 (x) as a perturbation which is smoothly turned on using a
parameter 0 ? t ? 1. We define f2 (x, t) to be a smooth interpolation between the trivial
f2 (x, t = 0) = 1 and the ?full? intractable f2 (x, t = 1) = f2 (x). Hence, we define
parametric densities and the corresponding free energy by p(x|t) = Z1t f1 (x)f2 (x, t) and
1
f1 (x)f2 (x, t) exp ?T g(x)
(15)
q(x|t) =
Zq (?, t)
n
o
?, t) = max ? ln Zq (?, t) + ?T ? .
Gq (?
(16)
?
D
E
we derive the following
Using eq. (11), and the fact that ? ln Z(?,t) = d ln f2 (x,t)
?t
dt
q(x|t)
?, t)
exact identity for the free energy G(?
?, 1) ? Gq (?
?, 0) = ?
Gq (?
Z
1
dt
0
d ln f2 (x, t)
dt
.
(17)
q(x|t)
?) = G(?
?, t = 1) and tractable
to relate the Gibbs free energy of the intractable model Gq (?
?, t = 0). An simple approximation is obtained for the case f2 (x, t) = [f2 (x)]t ,
model G(?
when the expectation over q(x|t) is replaced by the expectation over the tractable q(x|0),
ie
Z 1
d ln f2 (x, t)
?) ? G(?
?, 0) ?
?, 0) ? hln f2 (x)iq(x|0) . (18)
G(?
dt
= G(?
dt
0
q(x|0)
This result coincides with the variational approximation when we restrict the family of
optimizing functions to be of the form q(x|0).
4.1
Expectation Consistent Approximation
Our goal is to go beyond the variational approximation and capture more of the t dependence of the intractable part f2 in the term eq. (17). We will now use our assumption that
besides the family of distributions eq. (15), there is a second family which can be used as
an approximation to the distribution p(x|t). It is given by
1
r(x|t) =
f2 (x, t) exp ?T g(x) ,
(19)
Zr (?, t)
where the parameters ? will be chosen in such a way as to guarantee consistency for the
expectations of g, i.e. hg(x)ir(x|t) = ? . Using eq. (19) in place of q(x|t) in eq. (17), we
get the approximation
Z 1
d ln f2 (x, t)
?, 1) ? Gq (?
?, 0) ?
?, 1) ? Gr (?
?, 0) ,
Gq (?
dt
= Gr (?
(20)
dt
0
r(x|t)
where the last equality is derived from the fact that both types of densities eqs. (15) and
(19) contain the same exponential family. This allows us to carry out the integral over the
interaction strength t in eq. (20) in closed form without specifying the interpolating term
f2 (x, t) explicitly. Hence, the expectation consistent (EC) approximation is
?, 1) ? Gq (?
?, 0) + Gr (?
?, 1) ? Gr (?
?, 0) ? GEC (?
?) .
Gq (?
(21)
?, 0), Gr ? Gr (?
?, 1) and Gs ? Gr (?
?, 0) in
To simplify notation, we will write Gq ? Gq (?
the following.
5
Models with pair-wise interactions
Our framework works very naturally to the class of models eq. (3). The EC approximation eq. (21) will be based on approximating neglected correlations in the factorizing ap?, 1) and
proximation using a Gaussian distributions. The corresponding free energies G(?
?, 0) appearing in eq. (21) are thus found using the Gaussian free energy eq. (14) with J
r(?
and J = 0
1
(22)
GEC (m, M) = Gq (m, M, 0) ? mT Jm
2
1
1
+ max
ln det(? ? J) ? Tr ?(M ? mmT )
?
2
2
1
1
? max
ln det ? ? Tr ?(M ? mmT ) ,
?
2
2
where the free energy Gq (m, M, 0) will depend explicitly upon the potentials ?? (x? ).
6
Free energy minimization algorithms
In our approach, inference is based on the minimization of the free energy with respect to
its arguments ? . While the exact free energy is by construction a convex function in ? ,
our free energy approximation GEC = Gq + Gr ? Gs contains the concave contribution
?Gs and may not be convex. Hence, we may have potentially many local minima and
other stationary points, like saddlepoints. Moreover, the expression derived for G EC is
not a simple explicit functions of ? but contains additional optimizations over Lagrange
parameters. Consequently, we cannot expect that message-passing algorithms similar to
loopy belief propagation or expectation propagation (EP) [4] which sequentially update
moments and Lagrange parameters are guaranteed to converge.
6.1
Guaranteed convergence ? variational bounding
An iterative algorithms which is guaranteed to find at least stationary points of the Free
Energy is similar in spirit to the so-called double-loop approaches [9, 10]. The basic idea
is to minimize a decreasing sequence of convex upper bounds to GEC . Each convex upper
?) at the present iteration ? ? , i.e. usbound is derived by linearizing the concave term ?Gs (?
?
lbound
T
? ) ? Gs
?) = ?C? +?
? ?s , with C? ? ln Zq (??s ) and ??s = ?s (?
?? ). Since
ing Gs (?
(?
it is usually much easier to deal with the Lagrange parameters we convert the resulting
convex minimization problem into a concave maximization problem for the ??s
?) + Gr (?
?) ? ? T ??s + C?
? Gq (?
= min max ? ln Zq (?q ) ? ln Zr (?r ) + ? T (?q + ?r ? ??s ) + C?
? ? ,?
q
r
= max {? ln Zq (?q ) ? ln Zr (?r )|?q + ?r = ??s } + C?
?q ,?r
= max {? ln Zq (??s ? ?r ) ? ln Zr (?r ) + C? } .
(23)
?r
This can be summarized in the following double loop algorithm which is guaranteed not to
increase our free energy approximation.
?)
GEC (?
?) by
1. Outer loop: For fixed old value ? ? , bound the concave term ?Gs (?
?) go get the convex upper bound to GEC (?
?).
?Glbound
(?
s
2. Inner loop: Solve the concave maximization problem
max L
with L = ? ln Zq (??s ? ?r ) ? ln Zr (?r ) .
(24)
?r
Inserting the solution into ? (?r ) = hg(x)ir gives the new value ? ? for ? .
Currently, we either solve the non-linear inner-loop optimization by a sequential approach
that are computationally efficient when Gr is the free energy of a multivariate Gaussian or
by interior point methods [11, 12].
7
Simulations
We have tested the EC framework in a benchmark set-up proposed by Wainwright and
Jordan [12]. The stochastic variables are binary xi = ?1 with pair-wise couplings are
used. The N = 16 nodes are either fully connected or connected to nearest neighbors
in a 4-by-4 grid. The external field (observation) strengths ?i are drawn from a uniform
distribution ?i ? U[?dobs , dobs ] with dobs = 0.25. Three types of coupling strength
statistics are considered: repulsive (anti-ferromagnetic) Jij ? U[?2dcoup , 0], mixed Jij ?
U[?dcoup , +dcoup ] and attractive (ferromagnetic) Jij ? U[0,P
+2dcoup ] with dcoup > 0.
We compute the average one-norm error on the marginals:
i |p(xi = 1) ? p(xi =
1|Method)|/N , p(xi = 1) = (1 + mi )/2 over 100 trials testing the following Methods:
SP = sum-product (aka loopy belief propagation (BP) or Bethe approximation) and LD =
log-determinant maximization [12], EC factorized and EC structured. Results for SP and
LD are taken from Ref. [12]. For EC, we are minimizing the EC free energy eq. (22) where
Gq (m, M, 0) depend upon the approximation we are using. For the factorized model we
use the free energy eq. (12) and for the structured model we assume a single tractable
potential ?(x) in eq. (3) which contains all couplings on a spanning tree. For G q , we use
the free energy eq. (13). The spanning tree is defined by the following simple heuristic:
choose as next pair of nodes to link, the (so far unlinked) pair with strongest absolute
coupling |Jij | that will not cause a loop in the graph.
The results are summarized in table 1. The Bethe approximation always give inferior results compared to EC (note that only loopy BP convergent problem instances were used to
calculate the error [12]). This might be a bit surprising for the sparsely connected grids.
This indicates that loopy BP and too a lesser degree extensions building upon BP [5] are
only to be applied to really sparse graphs and/or weakly coupled nodes, where the error
induced by not using a properly normalized distribution can be expected to be small. We
also speculate that a structured variational approximation, using the same heuristics as described above to construct the spanning tree, in many cases will be superior to the Bethe
approximation as also observed by Ref. [5]. LD is a robust method which seems to be
limited in it?s achievable precision. EC structured is uniformly superior to all other approaches. Additional simulations (not included in the paper) also indicate that EC give
much improved estimates of free energies and two-node marginals when compared to the
Bethe- and Kikuchi-approximation.
8
Conclusion and outlook
We have introduced a novel method for approximate inference which tries to overcome
certain limitations of single approximating distributions by achieving consistency for two
of these on the same problem. While we have demonstrated its accuracy in this paper
only for a model with binary elements, it can also be applied to models with continuous
random variables or hybrid models with both discrete and continuous variables. We expect
that our method becomes most powerful when certain tractable substructures of variables
with strong dependencies can be identified in a model. Our approach would then allow
to deal well with the weaker dependencies between the groups. A generalization of our
method to treat graphical models beyond pair-wise interaction is obtained by iterating the
approximation. This is useful in cases, where an initial three term approximation G EC =
Table 1: The average one-norm error on marginals for the Wainwright-Jordan set-up.
Problem type
Graph
Full
Grid
Coupling
Repulsive
Repulsive
Mixed
Mixed
Attractive
Attractive
Repulsive
Repulsive
Mixed
Mixed
Attractive
Attractive
dcoup
0.25
0.50
0.25
0.50
0.06
0.12
1.0
2.0
1.0
2.0
1.0
2.0
SP
Mean
0.037
0.071
0.004
0.055
0.024
0.435
0.294
0.342
0.014
0.095
0.440
0.520
LD
Mean
0.020
0.018
0.020
0.021
0.027
0.033
0.047
0.041
0.016
0.038
0.047
0.042
Method
EC fac
Mean
0.003
0.031
0.002
0.022
0.004
0.117
0.153
0.198
0.011
0.082
0.125
0.177
EC struct
Mean
0.0017
0.0143
0.0013
0.0151
0.0031
0.0211
0.0031
0.0021
0.0018
0.0068
0.0028
0.0024
Gq + Gr ? Gs still contains non-tractable component free energies G.
References
[1] M. Opper and O. Winther, ?Gaussian processes for classification: Mean field algorithms,?
Neural Computation, vol. 12, pp. 2655?2684, 2000.
[2] M. Opper and O. Winther, ?Tractable approximations for probabilistic models: The adaptive
Thouless-Anderson-Palmer mean field approach,? Phys. Rev. Lett., vol. 86, pp. 3695, 2001.
[3] M. Opper and O. Winther, ?Adaptive and self-averaging Thouless-Anderson-Palmer mean field
theory for probabilistic modeling,? Phys. Rev. E, vol. 64, pp. 056131, 2001.
[4] T. P. Minka, ?Expectation propagation for approximate Bayesian inference,? in UAI 2001,
2001, pp. 362?369.
[5] T. Minka and Y. Qi, ?Tree-structured approximations by expectation propagation,? in NIPS 16,
S. Thrun, L. Saul, and B. Sch?olkopf, Eds. MIT Press, Cambridge, MA, 2004.
[6] Christopher M. Bishop, David Spiegelhalter, and John Winn, ?Vibes: A variational inference
engine for bayesian networks,? in Advances in Neural Information Processing Systems 15,
S. Thrun S. Becker and K. Obermayer, Eds., pp. 777?784. MIT Press, Cambridge, MA, 2003.
[7] H. Attias, ?A variational Bayesian framework for graphical models,? in Advances in Neural
Information Processing Systems 12, T. Leen et al., Ed. 2000, MIT Press, Cambridge.
[8] J. S. Yedidia, W. T. Freeman, and Y. Weiss, ?Generalized belief propagation,? in Advances in
Neural Information Processing Systems 13, T. K. Leen, T. G. Dietterich, and V. Tresp, Eds.,
2001, pp. 689?695.
[9] A. L. Yuille, ?CCCP algorithms to minimize the Bethe and Kikuchi free energies: convergent
alternatives to belief propagation,? Neural Comput., vol. 14, no. 7, pp. 1691?1722, 2002.
[10] T. Heskes, K. Albers, and H. Kappen, ?Approximate inference and constrained optimization,?
in UAI-03, San Francisco, CA, 2003, pp. 313?320, Morgan Kaufmann Publishers.
[11] S. Boyd and L. Vandenberghe, Convex Optimization, Cambridge University Press, 2004.
[12] M. J. Wainwright and M. I. Jordan, ?Semidefinite methods for approximate inference on graphs
with cycles,? Tech. Rep. UCB/CSD-03-1226, UC Berkeley CS Division, 2003.
| 2661 |@word trial:3 determinant:2 achievable:1 seems:3 norm:2 simulation:2 covariance:1 thereby:1 tr:3 outlook:1 ld:4 kappen:1 initial:1 carry:1 contains:4 moment:12 united:1 electronics:1 denoting:1 surprising:1 dx:1 john:1 partition:4 shape:1 update:1 stationary:2 short:1 manfred:1 node:13 gec:6 simpler:1 mathematical:1 constructed:1 introduce:1 expected:1 ica:1 isi:1 freeman:1 decreasing:1 jm:2 ihg:1 considering:1 becomes:2 dxf:1 notation:1 moreover:1 factorized:5 argmin:1 complimentary:1 minimizes:1 developed:3 differentiation:1 guarantee:1 berkeley:1 concave:5 exactly:1 uk:1 appear:1 understood:2 local:3 treat:2 interpolation:2 ap:1 might:3 specifying:1 relaxing:1 limited:1 palmer:2 bi:1 testing:1 bootstrap:1 matching:2 convenient:2 word:1 boyd:1 get:4 cannot:2 interior:1 optimize:1 demonstrated:1 go:2 convex:8 simplicity:1 deriving:1 vandenberghe:1 construction:2 exact:4 approximative:1 element:1 approximated:2 sparsely:1 persist:1 ep:3 observed:1 capture:1 calculate:1 ferromagnetic:2 connected:7 cycle:1 constrains:1 neglected:1 depend:2 weakly:1 yuille:1 upon:4 division:2 f2:21 completely:3 lbound:1 joint:1 fac:1 ole:1 neighborhood:1 heuristic:2 widely:1 solve:2 say:2 otherwise:1 statistic:1 gi:4 transform:1 sequence:1 propose:2 interaction:3 product:3 jij:6 gq:16 inserting:2 turned:1 combining:1 loop:6 adatap:1 olkopf:1 convergence:1 double:2 kikuchi:3 coupling:7 develop:2 ac:1 fixing:1 iq:4 derive:1 nearest:1 ij:3 school:1 albers:1 eq:20 strong:1 c:1 resemble:1 indicate:1 drawback:1 subsequently:1 stochastic:1 enable:1 require:1 f1:7 generalization:3 fix:1 really:1 adjusted:1 extension:1 mm:1 considered:2 plagued:1 exp:8 bj:1 mo:1 hgi:1 currently:1 successfully:1 minimization:3 mit:3 gaussian:13 always:2 factorizes:1 dcoup:6 encode:1 derived:3 properly:1 unsatisfactory:1 modelling:1 likelihood:1 indicates:1 aka:1 tech:1 contrast:1 helpful:1 inference:11 entire:1 dual:1 classification:1 development:1 constrained:2 uc:1 marginal:4 field:4 construct:2 eliminated:1 others:1 simplify:1 dg:1 divergence:4 densely:1 thouless:2 replaced:1 lebesgue:1 message:2 semidefinite:1 hg:7 edge:1 integral:4 necessary:2 tree:8 old:1 desired:1 instance:1 formalism:1 modeling:1 maximization:4 loopy:4 tractability:1 southampton:1 subset:2 uniform:1 gr:11 too:2 dependency:3 density:10 winther:4 ie:1 retain:1 probabilistic:7 informatics:1 connecting:1 together:1 again:1 central:1 choose:1 external:1 derivative:1 account:1 potential:4 speculate:1 summarized:2 explicitly:2 dobs:3 b2i:1 later:1 try:1 closed:1 linked:1 complicated:1 substructure:1 contribution:1 minimize:2 ni:2 ir:2 accuracy:1 unlinked:1 kaufmann:1 bayesian:3 strongest:1 phys:2 ed:4 definition:1 energy:36 pp:8 minka:3 naturally:1 mi:8 dxi:1 soton:1 treatment:1 dt:9 specify:1 improved:1 wei:1 formulation:1 evaluated:1 done:1 leen:2 anderson:2 stage:1 correlation:5 replacing:3 christopher:1 propagation:9 building:1 dietterich:1 effect:1 contain:1 true:3 normalized:1 hence:5 analytically:1 equality:1 deal:2 attractive:5 self:1 inferior:2 coincides:1 linearizing:1 generalized:2 prominent:1 criterion:1 ranging:1 variational:8 wise:3 novel:3 recently:1 superior:2 mt:3 tively:1 marginals:3 significant:1 cambridge:4 gibbs:6 ai:1 unconstrained:1 grid:5 consistency:6 heskes:1 hxi:2 gj:1 gt:1 closest:1 multivariate:4 optimizing:3 scenario:1 certain:5 binary:5 outperforming:1 rep:1 seen:1 minimum:1 additional:2 morgan:1 converge:1 full:3 smooth:1 technical:1 ing:1 characterized:1 cccp:1 va:5 qi:1 involving:1 basic:1 expectation:15 iteration:1 kernel:1 achieved:1 whereas:1 winn:1 publisher:1 sch:1 biased:1 induced:1 incorporates:1 spirit:1 jordan:3 call:1 ideal:1 enough:1 variety:3 xj:3 zi:2 restrict:1 identified:1 inner:2 idea:3 lesser:1 attias:1 det:3 expression:3 becker:1 passing:1 cause:1 useful:1 iterating:1 amount:1 cosh:1 generate:1 fulfilled:1 disjoint:1 write:2 discrete:1 vol:4 group:2 achieving:1 drawn:1 graph:10 sum:4 convert:1 powerful:1 communicate:1 place:1 family:7 mii:1 bit:1 bound:3 guaranteed:4 convergent:2 g:8 nontrivial:1 strength:3 constraint:2 bp:4 x2:1 aspect:1 argument:2 min:3 structured:6 developing:1 according:1 legendre:1 conjugate:2 rev:2 restricted:2 taken:1 lyngby:1 ln:30 computationally:1 tractable:15 repulsive:5 gaussians:2 operation:1 yedidia:1 apply:1 appearing:1 subtracted:1 alternative:2 struct:1 thomas:1 graphical:4 unifying:1 neglect:3 approximating:4 parametric:1 dependence:2 usual:1 so17:1 obermayer:1 link:2 separate:1 thrun:2 outer:1 trivial:2 spanning:4 denmark:2 besides:1 minimizing:2 kingdom:1 difficult:1 fe:1 potentially:1 relate:1 negative:1 zt:1 upper:3 observation:1 benchmark:2 anti:1 communication:1 rn:1 perturbation:1 arbitrary:1 hln:1 introduced:1 david:1 pair:5 kl:6 tap:2 engine:1 nip:1 beyond:2 usually:1 max:10 belief:4 wainwright:3 suitable:1 natural:1 hybrid:1 zr:5 scheme:1 x2i:1 spiegelhalter:1 dtu:1 coupled:1 tresp:1 relative:3 fully:2 expect:2 mixed:5 limitation:1 degree:1 consistent:5 intractability:2 summary:1 surprisingly:1 last:1 free:36 allow:3 weaker:1 neighbor:1 saul:1 absolute:1 sparse:1 overcome:1 opper:4 xn:1 calculated:2 lett:1 qualitatively:1 adaptive:4 san:1 ec:16 far:1 approximate:9 implicitly:2 keep:1 dealing:1 global:3 imm:1 proximation:1 sequentially:1 uai:2 francisco:1 factorize:1 xi:15 factorizing:4 continuous:4 iterative:1 zq:7 table:2 promising:1 bethe:7 reasonably:1 mj:2 robust:1 ca:1 interpolating:1 domain:1 sp:3 mmt:2 csd:1 owi:1 whole:1 bounding:1 ref:2 x1:1 fashion:1 precision:1 explicit:3 exponential:2 comput:1 bishop:1 dk:2 normalizing:1 intractable:10 sequential:1 importance:1 easier:1 vibe:1 entropy:1 smoothly:1 led:1 forming:1 lagrange:5 expressed:1 contained:1 mij:4 satisfies:1 ma:2 z1t:1 identity:1 goal:1 consequently:1 included:1 infinite:2 typical:1 uniformly:1 averaging:1 called:2 gij:2 ucb:1 support:1 arises:2 tested:1 |
1,830 | 2,662 | Methods Towards Invasive Human
Brain Computer Interfaces
Thomas Navin Lal1 , Thilo Hinterberger2 , Guido Widman3 ,
Michael Schr?oder4 , Jeremy Hill1 , Wolfgang Rosenstiel4 ,
Christian E. Elger3 , Bernhard Sch?olkopf1 and Niels Birbaumer2,5
Max-Planck-Institute for Biological Cybernetics, Tu? bingen, Germany
{navin,jez,bs}@tuebingen.mpg.de
2
Eberhard Karls University, Dept. of Medical Psychology and
Behavioral Neurobiology, T?ubingen, Germany
{thilo.hinterberger,niels.birbaumer}@uni-tuebingen.de
3
University of Bonn, Department of Epileptology, Bonn, Germany
{guido.widman,christian.elger}@ukb.uni-bonn.de
4
Eberhard Karls University, Dept. of Computer Engineering, Tu? bingen,
Germany {schroedm,rosenstiel}@informatik.uni-tuebingen.de
5
Center for Cognitive Neuroscience, University of Trento, Italy
1
Abstract
During the last ten years there has been growing interest in the development of Brain Computer Interfaces (BCIs). The field has mainly been
driven by the needs of completely paralyzed patients to communicate.
With a few exceptions, most human BCIs are based on extracranial electroencephalography (EEG). However, reported bit rates are still low. One
reason for this is the low signal-to-noise ratio of the EEG [16]. We are
currently investigating if BCIs based on electrocorticography (ECoG) are
a viable alternative. In this paper we present the method and examples
of intracranial EEG recordings of three epilepsy patients with electrode
grids placed on the motor cortex. The patients were asked to repeatedly imagine movements of two kinds, e.g., tongue or finger movements.
We analyze the classifiability of the data using Support Vector Machines
(SVMs) [18, 21] and Recursive Channel Elimination (RCE) [11].
1
Introduction
Completely paralyzed patients cannot communicate despite intact cognitive functions. The
disease Amyotrophic Lateral Sclerosis (ALS) for example, leads to complete paralysis of
the voluntary muscular system caused by the degeneration of the motor neurons. Birbaumer
et al. [1, 9] developed a Brain Computer Interface (BCI), called the Thought Translation
Device (TTD), which is used by several paralyzed patients. In order to use the interface,
patients have to learn to voluntary regulate their Slow Cortical Potentials (SCP). The system
then allows its users to write text on the screen of a computer or to surf the web. Although it
presents a major breakthrough, the system has two disadvantages. Not all patients manage
Figure 1: The left picture schematically shows the position of the 8x8 electrode grid of patient II. It was placed on the right hemisphere. As shown in the right picture the electrodes
are connected to the amplifier via cables that are passed through the skull.
to control their SCP. Furthermore the bit rate is quite low. A well-trained user requires
about 30 seconds to write one character.
Recently there has been increasing interest on EEG-based BCIs in the machine learning
community. In contrast to the TTD, in many BCI-systems the computer learns rather than
the system?s user [2, 5, 11]. Most such BCIs require a data collection phase during which
the subject repeatedly produces brain states of clearly separable locations. Machine learning techniques like Support Vector Machines or Fisher Discriminant are applied to the data
to derive a classifying function. This function can be used in online applications to identify
the different brain states produced by the subject.
The majority of BCIs is based on extracranial EEG-recordings during imagined limb
movements. We restrict ourselves to mentioning just a few publications [14, 15, 17, 22].
Movement-related cortical potentials in humans on the basis of electrocorticographical data
have also been studied, e.g. by [20]. Very recently the first work describing BCIs based on
electrocorticographic recordings was published [6, 13]. Successful approaches have been
developed using BCIs based on single unit, multiunit or field potentials recordings of primates. Serruya et al. taught monkeys to control a cursor on the basis of potentials from 7-30
motor cortex neurons [19]. The BCI developed by [3] enables monkeys to reach and grasp
using a robot arm. Their system is based on recordings from frontoparietal cell ensembles.
Driven by the success of BCIs for primates based on single unit or multiunit recordings,
we are currently developing a BCI-system that is based on ECoG recordings, as described
in the present paper.
2
Electrocorticography and Epilepsy
All patients presented suffer from a focal epilepsy. The epileptic focus - the part of the
brain which is responsible for the seizures - is removed by resection. Prior to surgery, the
epileptic focus has to be localized. In some complicated cases, this must be done by placing
electrodes onto the surface of the cortex as well as into deeper regions of the brain. The
skull over the region of interest is removed, the electrodes are positioned and the incision is
sutured. The electrodes are connected to a recording device via cables (cf. Figure 1). Over
a period of a 5 to 14 days ECoG is continuously recorded until the patient has had enough
seizures to precisely localize the focus [10]. Prior to surgery the parts of the cortex that are
covered by the electrodes are identified by the electric stimulation of electrodes.
In the current setup, the patients keep the electrode implants for one to two weeks. After
the implantation surgery, several days of recovery and follow-up examinations are needed.
Due to the tight time constraints, it is therefore not possible to run long experiments. Furthermore most of the patients cannot concentrate for a long period of time. Therefore only
a small amount of data could be collected.
Table 1: Positions of implanted electrodes. All three patients had an electrode grid implanted that partly covered the right or the left motor cortex.
patient
implanted electrodes
I
64-grid right hemisphere,
two 4-strip interhemisphere
64-grid right hemisphere
20-grid central,
four 16-strips frontal
II
III
3
task
trials
left vs. right hand
200
little left finger vs. tongue
little right finger vs. tongue
150
100
Experimental Situation and Data Acquisition
The experiments were performed in the department of epileptology of the University of
Bonn. We recorded ECoG data from three epileptic patients with a sampling rate of
1000Hz.
The electrode grids were placed on the cortex under the dura mater and covered the primary motor and premotor area as well as the fronto-temporal region either of the right or
left hemisphere. The grid-sizes ranged from 20 to 64 electrodes. Furthermore two of the
patients had additional electrodes implanted on other parts of the cortex (cf. Table 1). The
imagery tasks were chosen such that the involved parts of the brain
? were covered by the electrode grid
? were represented spatially separate in the primary motor cortex.
The expected well-localized signal in motor-related tasks suggested discrimination tasks
using imagination of hand, little finger, or tongue movements.
The patients were seated in a bed facing a monitor and were asked to repeatedly imagine
two different movements. At the beginning of each trial, a small fixation cross was displayed in the center of the screen. The 4 second imagination phase started with a cue that
was presented in the form of a picture showing either a tongue or a little finger for patients
II and III. The cue for patient I was an arrow pointing left or right. There was a short break
between the trials. The images which were used as a cue are shown in Figure 5.
4
Preprocessing
Starting half a second after the visualization of the task-cue, we extracted a window of
length 1.5 seconds from the data of each electrode. For every trial and every electrode we
thus obtained an EEG sequence that consisted of 1500 samples. The linear trend from every
sequence was removed. Following [8, 11, 15] we fitted a forward-backward autoregressive
model of order three to each sequence. The concatenated model parameters of the channels
together with the descriptor of the imagined task (i.e. +1, -1) form one training point. For
a given number n of EEG channels, a training point (x, y) is therefor a point in R 3n ?
{?1, 1}.
5
Channel Selection
The number of available training points is relatively small compared to the dimensionality
of the data. The data of patient III for example, consists of only 100 training points of
Figure 2: The patients were asked to repeatedly imagine two different movements that are
represented separately at the primary cortex, e.g. tongue and little finger movements. This
figure shows two stimuli that were used as a cue for imagery. The trial structure is shown
on the right. The imagination phase lasted four seconds. We extracted segments of 1.5
seconds from the ECoG recordings for the analysis.
dimension 252. This is a typical setting in which features selection methods can improve
classification accuracy.
Lal et al. [11] recently introduced a feature selection method for the special case of EEG
data. Their method is based on Recursive Feature Elimination (RFE) [7]. RFE is a backward feature selection method. Starting with the full data set, features are iteratively removed from the data until a stopping criteria is met. In each iteration a Support Vector
Machine (SVM) is trained and its weight vector is analyzed. The feature that corresponds
to the smallest weight vector entry is removed.
Recursive Channel Elimination (RCE) [11] treats features that belong to the data of a channel in a consistent way. As in RFE, in every iteration one SVM is trained. The evaluation
criteria that determines which of the remaining channels will be removed is the mean of the
weight vector entries that correspond to a channel?s features. All features of the channel
with the smallest mean value are removed from the data. The output of RCE is a list of
ranked channels.
6
Data Analysis
To begin with, we are interested in how well SVMs can learn from small ECoG data sets.
Furthermore we would like to understand how localized the classification-relevant information is, i.e. how many recording positions are necessary to obtain high classification
accuracy. We compare how well SVMs can generalize given the data of different subsets
of ECoG-channels:
(i) the complete data, i.e. all channels
(ii) the subset of channels suggested by RCE. In this setting we use the list of ranked
channels from RCE in the following way: For every l in the range of one to the
total number of channels, we calculate a 10-fold cross-validation error on the data
of the l best-ranked channels. We use the subset of channels which leads to the
lowest error estimate.
(iii) the two best-ranked channels by RCE. The underlying assumption used here is
that the classification-relevant information is extremely localized and that two correctly chosen channels contain sufficient information for classification purposes.
(iv) two channels drawn at random.
Throughout the paper we use linear SVMs. For regularization purposes we use a ridge on
the kernel matrix which corresponds to a 2-norm penalty on the slack variables [4].
C4
muV
C3
C2
C1
0
500
1000
1500
time [ms]
2000
2500
3000
3500
Figure 3: This plot shows ECoG recordings from 4 channels while the patient was imagining movements. The distance of two horizontal lines decodes 100?V . The amplitude of
the recordings ranges roughly from -100 ?V to +100 ?V which is on the order of five to
ten times the amplitude measured with extracranial EEG.
To evaluate the classification performance of an SVM that is trained on a specific subset
of channels we calculate its prediction error on a separate test set. We use a double-crossvalidation scheme - the following procedure is repeated 50 times:
We randomly split the data into a training set (80%) and a test set (20%). Via 10-fold
cross-validation on the training set we estimate all parameters for the different considered
subsets (i)-(iv):
(i) The ridge is estimated.
(ii) On the basis of the training set RCE suggests a subset of channels. We restrict the
training set as well as the test set to these channels. A ridge-value is then estimated
from the restricted training set.
(iii) We restrict the training set and the test set to the 2 best ranked channels by RCE.
The ridge is then estimated on the restricted training set.
(iv) The ridge is estimated.
We then train an SVM on the (restricted) training set using the estimated ridge. The trained
model is tested on the (restricted) test set. For (i)-(iv) we obtain 50 test error estimates from
the 50 repetitions for each patient. Table 2 summarizes the results.
7
Results
The results in Table 2 show that the generalization ability can significantly be increased by
RCE. For patient I the error decreases from 38% to 24% when using the channel subsets
suggested by RCE. In average RCE selects channel subsets of size 5.8. For patient II the
number of channels is reduced to one third but the channel selection process does not yield
an increased accuracy. The error of 40% can be reduced to 23% for patient III using in
average 5 channels selected by RCE.
For patients I and III the choice of the best 2 ranked channels leads to a much lower error
as well. The direct comparison of the results using the two best ranked channels to two
randomly chosen channels shows how well the RCE ranking method works: For patient
three the error drops from chance level for two random channels to 18 % using the two
best-ranked channels.
The reason why there is such a big difference in performance for patient III when comparing (i) and (iii) might be, that out of the 84 electrodes, only 20 are located over or close to
the motor cortex. RCE successfully identifies the important electrodes.
In contrast to patient III, the electrodes of patient II are all more or less located close to
Table 2: Classification Results. We compare the classification accuracy of SVMs trained
on the data of different channel subsets: (i) all ECoG-channels, (ii) the subset determined
by Recursive Channel Elimination (RCE), (iii) the subset consisting of the two best ranked
channels by RCE and (iv) two randomly drawn channels. The mean errors of 50 repetitions
are given along with the standard deviations. The test error can significantly be reduced by
RCE for two of the three patients. Using the two best ranked channels by RCE also yields
good results for two patients. SVMs trained on two random channels show performance
better than chance only for patient II.
pat
I
II
III
all channels (i)
#channels error
74
64
84
0.382 ? 0.071
0.257 ?0.076
0.4 ?0.1
RCE cross-val. (ii)
#channels error
5.8
21.5
5.0
0.243 ? 0.063
0.268 ? 0.080
0.233 ?0.13
RCE top 2 (iii)
error
random 2 (iv)
error
0.244 ? 0.078
0.309 ? 0.086
0.175 ? 0.078
chance level
0.419 ? 0.123
chance level
the motor cortex. This explains why data from two randomly drawn channels can yield
a classification rate better than chance. Furthermore patient II had the fewest electrodes
implanted and thus the chance of randomly choosing an electrode close to an important
location is higher than for the other two patients.
8
Discussion
We recorded ECoG-data from three epilepsy patients during a motor imagery experiment.
Although only few data were collected, the following conclusions can be drawn:
? The data of all three patients is reasonably well classifiable. The error rates range
from 17.5% to 23.3%. This is still high compared to the best error rates from BCI
based on extracranial EEG which are as low as 10% (e.g. [12]). Please note that
we used 1.5 seconds data from each trial only and that very few training points
(100-200) were available. Furthermore, extracranial EEG has been studied and
developed for a number of years.
? Recursive Channel Elimination (RCE) shows very good performance. RCE successfully identifies subsets of ECoG-channels that lead to good classification performance. On average, RCE leads to a significantly improved classification rate
compared to a classifier that is based on the data of all available channels.
? Poor classification rates using two randomly drawn channels and high classification rates using the two best-ranked channels by RCE suggest that classification
relevant information is focused on small parts of the cortex and depends on the
location of the physiological function.
? The best ranked RCE-channels correspond well with the results from the electric
stimulation (cf. Figure 8).
9
Ongoing Work and Further Research
Although our preliminary results indicate that invasive Brain Computer Interfaces may
be feasible, a number of questions need to be investigated in further experiments. For
instance, it is still an open question whether the patients are able to adjust to a trained
classifier and whether the classifying function can be transferred from session to session.
Moreover, experiments that are based on tasks different from motor imaginary need to
X X X
X X
X
X X X
Figure 4: Electric stimulation of the implanted electrodes helps to identify the parts of the
cortex that are covered by the electrode grid. This information is necessary for the surgery.
The red (solid) dots on the left picture mark the motor cortex of patient II as identified
by the electric stimulation method. The positions marked with yellow crosses correspond
to the epileptic focus. The red points on the right image are the best ranked channels by
Recursive Channel Elimination (RCE). The RCE-channels correspond well to the results
from the electro stimulation diagnosis.
be implemented and tested. It is quite conceivable that the tasks that have been found to
work well for extracranial EEG are not ideal for ECoG. Likewise, it is unclear whether our
preprocessing and machine learning methods, originally developed for extracranial EEG
data, are well adapted to the different type of data that ECoG delivers.
Acknowledgements
This work was supported in part by the Deutsche Forschungsgemeinschaft (SFB 550, B5
and grant RO 1030/12), by the National Institute of Health (D.31.03765.2), and by the
IST Programme of the European Community, under the PASCAL Network of Excellence,
IST-2002-506778. T.N.L. was supported by a grant from the Studienstiftung des deutschen
Volkes. Special thanks go to Theresa Cooke.
References
[1] N. Birbaumer, N. Ghanayim, T. Hinterberger, I. Iversen, B. Kotchoubey, A. K u? bler,
J. Perelmouter, E. Taub, and H. Flor. A spelling device for the paralysed. Nature,
398:297?298, 1999.
[2] B. Blankertz, G. Curio, and K. M?uller. Classifying single trial EEG: Towards brain
computer interfacing. In T.K. Leen, T.G. Dietterich, and V. Tresp, editors, Advances
in Neural Information Processing Systems, volume 14, Cambridge, MA, USA, 2001.
MIT Press.
[3] J.M. Carmena, M.A Lebedev, R.E Crist, J.E O?Doherty, D.M. Santucci, D. Dimitrov,
P.G. Patil, C.S Henriquez, and M.A. Nicolelis. Learning to control a brain-machine
interface for reaching and grasping by primates. PLoS Biology, 1(2), 2003.
[4] C. Cortes and V. Vapnik. Support-vector networks. Machine Learning, 20:273?297,
1995.
[5] J. del R. Millan, F. Renkens, J. Mourino, and W. Gerstner. Noninvasive brain-actuated
control of a mobile robot by human eeg. IEEE Transactions on Biomedical Engineering. Special Issue on Brain-Computer Interfaces, 51(6):1026?1033, June 2004.
[6] B. Graimann, J. E. Huggins, S. P. Levine, and G. Pfurtscheller. Towards a direct brain
interface based on human subdural recordings and wavelet packet analysis. IEEE
Trans. IEEE Transactions on Biomedical Engineering, 51(6):954?962, 2004.
[7] I. Guyon, J. Weston, S. Barnhill, and V. Vapnik. Gene selection for cancer classification using support vector machines. Journal of Machine Learning Research,
3:1439?1461, March 2003.
[8] S. Haykin. Adaptive Filter Theory. Prentice-Hall International, Inc., Upper Saddle
River, NJ, USA, 1996.
[9] T. Hinterberger, J. Kaiser, A. Kbler, N. Neumann, and N. Birbaumer. The Thought
Translation Device and its Applications to the Completely Paralyzed. In Diebner,
Druckrey, and Weibel, editors, Sciences of the Interfaces. Genista-Verlag, T u? bingen,
2001.
[10] J. Engel Jr. Presurgical evaluation protocols. In Surgical Treatment of the Epilepsies,
pages 740?742. Raven Press Ltd., New York, 2nd edition, 1993.
[11] T.N. Lal, M. Schr?oder, T. Hinterberger, J. Weston, M. Bogdan, N. Birbaumer, and
B. Sch?olkopf. Support Vector Channel Selection in BCI. IEEE Transactions on
Biomedical Engineering. Special Issue on Brain-Computer Interfaces, 51(6):1003?
1010, June 2004.
[12] S. Lemm, C. Sch?afer, and G. Curio. BCI Competition 2003 - Data Set III: Probabilistic Modeling of Sensorimotor mu-Rhythms for Classification of Imaginary Hand
Movements. IEEE Transactions on Biomedical Engineering. Special Issue on BrainComputer Interfaces, 51(6):1077?1080, June 2004.
[13] E. C. Leuthardt, G. Schalk, J. R. Wolpaw, J. G. Ojemann, and D. W. Moran. A
braincomputer interface using electrocorticographic signals in humans. Journal of
Neural Engineering, 1:63?71, 2004.
[14] D.J. McFarland, L.M. McCane, S.V. David, and J.R. Wolpaw. Spatial filter selection
for EEG-based communication. Electroencephalography and Clinical Neurophysiology, 103:386?394, 1997.
[15] G. Pfurtscheller., C. Neuper amd A. Schlo? gl, and K. Lugger. Separability of EEG
signals recorded during right and left motor imagery using adaptive autoregressive
parameters. IEEE Transactions on Rehabilitation Engineering, 6(3):316?325, 1998.
[16] J. Raethjen, M. Lindemann, M. D?umpelmann, R. Wenzelburger, H. Stolze, G. Pfister,
C. E. Elger, J. Timmer, and G. Deuschl. Corticomuscular coherence in the 6-15 hz
band: is the cortex involved in the generation of physiologic tremor? Experimental
Brain Research, 142:32?40, 2002.
[17] H. Ramoser, J. M?uller-Gerking, and G. Pfurtscheller. Optimal spatial filtering of single trial EEG during imagined hand movement. IEEE Transactions on Rehabilitation
Engineering, 8(4):441?446, 2000.
[18] B. Sch?olkopf and A. Smola. Learning with Kernels. MIT Press, Cambridge, USA,
2002.
[19] M.D. Serruya, N.G Hatsopoulos, L. Paninski, M.R. Fellows, and Donoghue J.P. Instant neural control of a movement signal. Nature, 416:141?142, 2002.
[20] C. Toro, G. Deuschl, R. Thatcher, S. Sato, C. Kufta, and M. Hallett. Event-related
desynchronization and movement-related cortical potentials on the ECoG and EEG.
Electroencephalography Clinical Neurophysiology, 5:380?389, 1994.
[21] V. N. Vapnik. Statistical Learning Theory. John Wiley and Sons, New York, USA,
1998.
[22] R. Wolpaw and D.J McFarland. Multichannel EEG-based brain-computer communication. Electroencephalography and Clinical Neurophysiology, 90:444?449, 1994.
| 2662 |@word neurophysiology:3 trial:8 norm:1 nd:1 open:1 solid:1 imaginary:2 current:1 comparing:1 must:1 john:1 toro:1 christian:2 motor:13 enables:1 plot:1 olkopf1:1 drop:1 v:3 discrimination:1 cue:5 half:1 device:4 selected:1 beginning:1 short:1 haykin:1 location:3 five:1 along:1 c2:1 direct:2 viable:1 consists:1 fixation:1 behavioral:1 classifiability:1 excellence:1 expected:1 roughly:1 mpg:1 growing:1 brain:17 little:5 window:1 electroencephalography:4 increasing:1 begin:1 underlying:1 moreover:1 deutsche:1 lowest:1 kind:1 monkey:2 developed:5 sutured:1 nj:1 temporal:1 fellow:1 every:5 ro:1 classifier:2 control:5 unit:2 medical:1 grant:2 renkens:1 planck:1 timmer:1 engineering:8 treat:1 despite:1 crist:1 might:1 studied:2 suggests:1 mentioning:1 range:3 responsible:1 recursive:6 wolpaw:3 procedure:1 area:1 thought:2 significantly:3 suggest:1 cannot:2 onto:1 selection:8 close:3 prentice:1 center:2 go:1 starting:2 tremor:1 focused:1 recovery:1 amyotrophic:1 imagine:3 user:3 guido:2 trend:1 located:2 electrocorticographic:2 deuschl:2 levine:1 calculate:2 degeneration:1 region:3 connected:2 grasping:1 movement:13 removed:7 decrease:1 plo:1 hatsopoulos:1 disease:1 mu:1 asked:3 ojemann:1 electrocorticography:2 ukb:1 trained:8 tight:1 segment:1 completely:3 basis:3 represented:2 finger:6 fewest:1 train:1 choosing:1 quite:2 premotor:1 bci:7 ability:1 bler:1 online:1 sequence:3 tu:2 relevant:3 bed:1 competition:1 olkopf:2 trento:1 rfe:3 crossvalidation:1 carmena:1 electrode:25 double:1 neumann:1 produce:1 help:1 derive:1 bogdan:1 measured:1 implemented:1 indicate:1 met:1 concentrate:1 filter:2 human:6 packet:1 elimination:6 explains:1 require:1 generalization:1 preliminary:1 biological:1 ecog:14 considered:1 hall:1 week:1 pointing:1 major:1 smallest:2 purpose:2 niels:2 currently:2 repetition:2 engel:1 successfully:2 uller:2 mit:2 clearly:1 interfacing:1 rather:1 reaching:1 frontoparietal:1 mobile:1 publication:1 focus:4 june:3 mainly:1 lasted:1 contrast:2 hill1:1 stopping:1 interested:1 germany:4 selects:1 issue:3 classification:16 pascal:1 elger:2 development:1 spatial:2 breakthrough:1 special:5 field:2 sampling:1 biology:1 placing:1 schroedm:1 implantation:1 stimulus:1 few:4 mater:1 randomly:6 national:1 phase:3 ourselves:1 consisting:1 volkes:1 resection:1 amplifier:1 interest:3 evaluation:2 grasp:1 adjust:1 analyzed:1 paralysed:1 necessary:2 iv:6 fronto:1 fitted:1 tongue:6 increased:2 instance:1 modeling:1 disadvantage:1 deviation:1 entry:2 subset:12 successful:1 reported:1 perelmouter:1 muv:1 thanks:1 gerking:1 eberhard:2 international:1 river:1 ghanayim:1 probabilistic:1 michael:1 together:1 continuously:1 lebedev:1 imagery:4 central:1 recorded:4 manage:1 hinterberger:4 cognitive:2 imagination:3 jeremy:1 potential:5 de:5 inc:1 caused:1 ranking:1 depends:1 performed:1 break:1 wolfgang:1 analyze:1 red:2 complicated:1 accuracy:4 descriptor:1 likewise:1 ensemble:1 correspond:4 identify:2 yield:3 surgical:1 yellow:1 generalize:1 decodes:1 schlo:1 produced:1 informatik:1 kotchoubey:1 cybernetics:1 published:1 subdural:1 reach:1 barnhill:1 strip:2 acquisition:1 sensorimotor:1 involved:2 invasive:2 treatment:1 dimensionality:1 amplitude:2 positioned:1 higher:1 originally:1 day:2 follow:1 improved:1 leen:1 done:1 furthermore:6 just:1 biomedical:4 smola:1 jez:1 widman:1 until:2 hand:4 horizontal:1 navin:2 web:1 del:1 lindemann:1 bcis:9 usa:4 dietterich:1 ranged:1 rosenstiel:1 consisted:1 contain:1 regularization:1 spatially:1 iteratively:1 during:6 please:1 rhythm:1 criterion:2 m:1 complete:2 ridge:6 doherty:1 delivers:1 interface:12 karls:2 image:2 recently:3 stimulation:5 birbaumer:5 volume:1 imagined:3 belong:1 epilepsy:5 taub:1 cambridge:2 grid:10 focal:1 session:2 therefor:1 had:4 dot:1 robot:2 afer:1 cortex:15 surface:1 intracranial:1 italy:1 hemisphere:4 driven:2 verlag:1 ubingen:1 success:1 additional:1 period:2 paralyzed:4 signal:5 ii:13 full:1 cross:5 long:2 clinical:3 dept:2 prediction:1 implanted:6 patient:40 iteration:2 kernel:2 serruya:2 cell:1 c1:1 schematically:1 separately:1 dimitrov:1 sch:4 flor:1 recording:13 subject:2 hz:2 electro:1 ideal:1 iii:14 enough:1 split:1 forschungsgemeinschaft:1 psychology:1 restrict:3 identified:2 donoghue:1 whether:3 epileptic:4 b5:1 passed:1 sfb:1 ltd:1 penalty:1 suffer:1 bingen:3 york:2 repeatedly:4 oder:1 physiologic:1 covered:5 amount:1 ten:2 band:1 svms:6 hallett:1 reduced:3 multichannel:1 neuroscience:1 estimated:5 correctly:1 diagnosis:1 write:2 taught:1 ist:2 four:2 monitor:1 localize:1 drawn:5 backward:2 year:2 run:1 communicate:2 classifiable:1 throughout:1 guyon:1 coherence:1 summarizes:1 seizure:2 bit:2 fold:2 sato:1 adapted:1 precisely:1 constraint:1 lemm:1 bonn:4 ttd:2 extremely:1 separable:1 relatively:1 transferred:1 department:2 developing:1 march:1 poor:1 sclerosis:1 jr:1 son:1 character:1 separability:1 skull:2 cable:2 b:1 primate:3 rehabilitation:2 huggins:1 restricted:4 visualization:1 describing:1 slack:1 needed:1 available:3 limb:1 regulate:1 alternative:1 thomas:1 top:1 remaining:1 cf:3 patil:1 schalk:1 rce:27 iversen:1 instant:1 concatenated:1 surgery:4 question:2 kaiser:1 primary:3 spelling:1 unclear:1 conceivable:1 distance:1 separate:2 lateral:1 majority:1 amd:1 collected:2 tuebingen:3 discriminant:1 reason:2 length:1 ratio:1 setup:1 multiunit:2 upper:1 neuron:2 displayed:1 pat:1 voluntary:2 situation:1 neurobiology:1 communication:2 schr:2 community:2 introduced:1 david:1 c3:1 lal:2 c4:1 trans:1 able:1 suggested:3 mcfarland:2 max:1 event:1 ranked:13 examination:1 nicolelis:1 braincomputer:2 santucci:1 arm:1 blankertz:1 scheme:1 improve:1 picture:4 identifies:2 started:1 x8:1 health:1 tresp:1 text:1 prior:2 weibel:1 acknowledgement:1 val:1 generation:1 filtering:1 facing:1 localized:4 validation:2 sufficient:1 consistent:1 editor:2 classifying:3 seated:1 translation:2 cooke:1 cancer:1 deutschen:1 placed:3 last:1 supported:2 gl:1 deeper:1 understand:1 institute:2 dimension:1 cortical:3 noninvasive:1 autoregressive:2 forward:1 collection:1 adaptive:2 preprocessing:2 programme:1 transaction:6 uni:3 bernhard:1 keep:1 gene:1 investigating:1 paralysis:1 why:2 table:5 channel:56 learn:2 reasonably:1 nature:2 henriquez:1 actuated:1 eeg:20 imagining:1 investigated:1 european:1 gerstner:1 electric:4 protocol:1 ramoser:1 surf:1 arrow:1 big:1 noise:1 edition:1 repeated:1 screen:2 slow:1 wiley:1 pfurtscheller:3 position:4 leuthardt:1 third:1 learns:1 wavelet:1 specific:1 showing:1 moran:1 list:2 desynchronization:1 svm:4 physiological:1 cortes:1 raven:1 studienstiftung:1 curio:2 vapnik:3 implant:1 cursor:1 paninski:1 saddle:1 thilo:2 corresponds:2 determines:1 chance:6 extracted:2 ma:1 weston:2 marked:1 towards:3 fisher:1 feasible:1 muscular:1 typical:1 determined:1 called:1 total:1 pfister:1 partly:1 experimental:2 intact:1 neuper:1 exception:1 millan:1 support:6 scp:2 mark:1 theresa:1 frontal:1 ongoing:1 evaluate:1 tested:2 |
1,831 | 2,663 | Contextual models for object detection using
boosted random fields
Antonio Torralba
MIT, CSAIL
Cambridge, MA 02139
[email protected]
Kevin P. Murphy
UBC, CS
Vancouver, BC V6T 1Z4
[email protected]
William T. Freeman
MIT, CSAIL
Cambridge, MA 02139
[email protected]
Abstract
We seek to both detect and segment objects in images. To exploit both local image data as well as contextual information, we introduce Boosted
Random Fields (BRFs), which uses Boosting to learn the graph structure and local evidence of a conditional random field (CRF). The graph
structure is learned by assembling graph fragments in an additive model.
The connections between individual pixels are not very informative, but
by using dense graphs, we can pool information from large regions of
the image; dense models also support efficient inference. We show how
contextual information from other objects can improve detection performance, both in terms of accuracy and speed, by using a computational
cascade. We apply our system to detect stuff and things in office and
street scenes.
1
Introduction
Our long-term goal is to build a vision system that can examine an image and describe what
objects are in it, and where. In many images, such as Fig. 5(a), objects of interest, such as
the keyboard or mouse, are so small that they are impossible to detect just by using local
features. Seeing a blob next to a keyboard, humans can infer it is likely to be a mouse; we
want to give a computer the same abilities.
There are several pieces of related work. Murphy et al [9] used global scene context to
help object recognition, but did not model relationships between objects. Fink and Perona
[4] exploited local dependencies in a boosting framework, but did not allow for multiple
rounds of communication between correlated objects. He et al [6] do not model connections
between objects directly, but rather they induce such correlations indirectly, via a bank of
hidden variables, using a ?restricted Boltzmann machine? architecture.
In this paper, we exploit contextual correlations between the object classes by introducing
Boosted Random Fields (BRFs). Boosted random fields build on both boosting [5, 10]
and conditional random fields (CRFs) [8, 7, 6]. Boosting is a simple way of sequentially
constructing ?strong? classifiers from ?weak? components, and has been used for singleclass object detection with great success [12]. Dietterich et al [3] combine boosting and
1D CRFs, but they only consider the problem of learning the local evidence potentials; we
consider the much harder problem of learning the structure of a 2D CRF.
Standard applications of MRFs/ CRFs to images [7] assume a 4-nearest neighbor grid
structure. While successful in low-level vision, this structure will fail in capturing important long distance dependencies between whole regions and across classes. We propose
a method for learning densely connected random fields with long range connections. The
topology of these connections is chosen by a weak learner which has access to a library
of graph fragments, derived from patches of labeled training images, which reflect typical
spatial arrangments of objects (similar to the segmentation fragments in [2]). At each round
of the learning algorithm, we add more connections from other locations in the image and
from other classes (detectors). The connections are assumed to be spatially invariant, which
means this update can be performed using convolution followed by a sigmoid nonlinearity.
The resulting architecture is similar to a convolutional neural network, although we used a
stagewise training procedure, which is much faster than back propagation.
In addition to recognizing things, such as cars and people, we are also interested in recognizing spatially extended ?stuff? [1], such as roads and buildings. The traditional sliding
window approach to object detection does not work well for detecting ?stuff?. Instead, we
combine object detection and image segmentation (c.f., [2]) by labeling every pixel in the
image. We do not rely on a bottom-up image segmentation algorithm, which can be fragile
without top-down guidance.
2
Learning potentials and graph structure
A conditional random field (CRF) is a distribution of the form
Y
1 Y
?i (Si )
?i,j (Si , Sj )
P (S|x) =
Z i
j?Ni
where x is the input (e.g., image), Ni are the neighbors of node i, and Si are labels. We
have assumed pairwise potentials for notational simplicity. Our goal is to learn the local
evidence potentials, ?i , the compatibility potentials ?, and the set of neighbors Ni .
We propose the following simple approximation: use belief propagation (BP) to estimate
the marginals, P (Si |x), and then use boosting to maximize the likelihood of each node?s
training data with respect to ?i and ?.
In more detail, the algorithm is as follows. At iteration t, the goal is to minimize the
negative log-likelihood of the training data. As in [11], we consider the per-label loss (i.e.,
we use marginal probabilities), as opposed to requiring that the joint labeling be correct (as
in Viterbi decoding). Hence the cost function to be minimized is
Y
YY
YY
?
?
Jt =
Jit = ?
bti,m (Si,m ) = ?
bti,m (+1)Si,m bti,m (?1)1?Si,m (1)
m
i
m
i
i
?
= (Si,m +
where Si,m ? {?1, +1} is the true label for pixel i in training case m, Si,m
t
1)/2 ? {0, 1} is just a relabeling, and bi,m = [P (Si = ?1|xm , t), P (Si = 1|xm , t)] is the
belief state at node i given input image xm after t iterations of the algorithm.
The belief at node i is given by the following (dropping the dependence on case m)
bti (?1) ? ?ti (?1) Mit (?1) where Mit is the product of all the messages coming into i
from all its neighbors at time t and where the message that k sends to i is given by
Y
X
btk (sk )
t+1
(2)
Mit+1 (?1) =
?t+1
(?1)
?
(?1)
=
?
(s
,
?1)
k,i
k
k?i
k?i
?ti?k (sk )
k?Ni
sk ?{?1,+1}
where ?k,i is the compatility between nodes k and i. If we assume that the local potentials
t
t
have the form ?ti (si ) = [eFi /2 ; e?Fi /2 ], where Fit is some function of the input data, then:
bti (+1) = ?(Fit + Gti ),
Gti = log Mit (+1) ? log Mit (?1)
(3)
?u
where ?(u) = 1/(1 + e ) is the sigmoid function. Hence each term in Eq. 1 simplifies to
a cost function similar to that used in boosting:
X
t
t
log Jit =
log 1 + e?Si,m (Fi,m +Gi,m ) .
(4)
m
1. Input: a set of labeled pairs {xi,m ; Si,m }, bound T
Output: Local evidence functions fit (x) and message update functions git (bNi ).
t=0
t=0
2. Initialize: bt=0
i,m = 0; Fi,m = 0; Gi,m = 0
3. For t=1..T.
(a) Fit local potential fi (xi,m ) by weighted LS to
t
t
t
Yi,m
= Si,m (1 + e?Si,m (Fi +Gi,m ) )
t
. compatibilities git (bt?1
(b) Fit
Ni ,m ) to Yi,m by weighted LS.
t?1
t
(c) Compute local potential Fi,m
= Fi,m
+ fit (xi,m )
(e) Update
(f) Update
Pt
g n (bt?1
Ni ,m )
n=1 i
t
t
t
the beliefs bi,m = ?(Fi,m + Gi,m )
t+1
weights wi,m
= bti,m (?1) bti,m (+1)
(d) Compute compatibilities Gti,m =
Figure 1: BRF training algorithm.
We assume that the graph is very densely connected so that the information that
one single node sends to another is so small that we can make the approximation
t+1
?t+1
k?i (+1)/?k?i (?1) ? 1. (This is a reasonable approximation in the case of images,
where each node represents a single pixel; only when the influence of many pixels is taken
into account will the messages become informative.) Hence
P
btk,m (sk )
sk ?[?1,+1] ?k,i (sk , +1) ?ti?k (sk )
Mit+1 (+1) X
t+1
Gi = log t+1
=
log P
(5)
btk,m (sk )
Mi (?1)
?
(s
,
?1)
k
t
sk ?[?1,+1] k,i k
?i?k (sk )
P
t
X
s ?[?1,+1] ?k,i (sk , +1) bk,m (sk )
?
log P k
(6)
t
sk ?[?1,+1] ?k,i (sk , ?1) bk,m (sk )
k
With this simplification,
is now a non-linear function of the beliefs Gt+1
(~btm ) at
i
iteration t. Therefore, We can write the beliefs at iteration t as a function of the local
evidences and the beliefs at time t ? 1: bti (+1) = ?(Fit (xi,m ) + Gti (~bt?1
m )). The key idea
behind BRFs is to use boosting to learn the G functions, which approximately implement
message passing in densely connected graphs. We explain this in more detail below.
Gt+1
i
2.1
Learning local evidence potentials
Defining Fit (xi,m ) = Fit?1 (xi,m ) + fit (xi,m ) as an additive model, where xi,m are the
features of training sample m at node i, we can learn this function in a stagewise fashion
by optimizing the second order Taylor expansion of Eq. 4 wrt fit , as in logitBoost [5]:
X
t
t
t
arg min
log
J
?
arg
min
wi,m
(Yi,m
? fit (xi,m ))2
(7)
i
t
t
fi
fi
m
?Si,m (Fit +Gti,m )
t
where Yi,m
= Si,m (1+e
). In the case that the weak learner is a ?regression
stump?, fi (x) = ah(x)+b, we can find the optimal a, b by solving a weighted least squares
t
problem, with weights wi,m
= bti (?1) bti (+1); we can find the best basis function h(x) by
searching over all elements of a dictionary.
2.2
Learning compatibility potentials and graph structure
In this section, we discuss how to learn the compatibility functions ?ij , and hence the
structure of the graph. Instead of learning the compatibility functions ?ij , we propose to
1. Input: a set of inputs {xi,m } and functions fit , git
Output: Set of beliefs bi,m and MAP estimates Si,m .
t=0
t=0
2. Initialize: bt=0
i,m = 0; Fi,m = 0; Gi,m = 0
3. From t = 1 to T , repeat
t?1
t
(a) Update local evidences Fi,m
= Fi,m
+ fit (xi,m )
(b) Update compatibilities Gti,m =
(c) Compute current beliefs
bti,m
=
Pt
g n (bt?1
Ni ,m )
n=1 i
t
?(Fi,m + Gti,m )
4. Output classification is Si,m = ? bti,m > 0.5
Figure 2: BRF run-time inference algorithm.
t+1
learn directly the function Gt+1
as we
iP . We propose to use an additive model for Gi
t
t+1
n ~t
t
did for learning F : Gi,m = n=1 gi (bm ), where ~bm is a vector with the beliefs of all
nodes in the graph at iteration t for the training sample m. The weak learners gin (~btm ) can
be regression stumps with the form gin (~btm ) = a?(w
~ ? ~btm > ?) + b, where a, b, ? are the
parameters of the regression stump, and w
~ i is a set of weights selected from a dictionary.
In the case of a graph with weak and almost symmetrical connections (which holds if
?(s1 , s2 ) ? 1, for all (s1 , s2 ), which implies the messages are not very informative) we can
further simplify the function Gt+1
by approximating it as a linear function of the beliefs:
i
X
Gt+1
?k,i btk,m (+1) + ?k,i
(8)
i,m =
k?Ni
This step reduces the computational cost. The weak learners gin (~btm ) will also be linear
t
functions. Hence the belief update simplifies to bt+1
?i ?~btm + ?i + Fi,m
), which
i,m (+1) = ?(~
is similar to the mean-field update equations. The neighborhood Ni over which we sum
incoming messages is determined by the graph structure, which is encoded in the non-zero
values of ?i . Each weak learner gin will compute a weighted combination of the beliefs of
the some subset of the nodes; this subset may change from iteration to iteration, and can be
quite large. At iteration t, we choose the weak learner git so as to minimize
Pt?1 n t?1
X
t
t t?1
log Jit (bt?1 ) = ?
log 1 + e?Si,m (Fi,m +gi (bm )+ n=1 gi (bm ))
m
which reduces to a weighted least squares problem similar to Eq. 7. See Fig. 1 for the
pseudo-code for the complete learning algorithm, and Fig. 2 for the pseudo-code for runtime inference.
3
BRFs for multiclass object detection and segmentation
With the BRF training algorithm in hand, we describe our approach for multiclass object
detection and region-labeling using densely connected BRFs.
3.1
Weak learners for detecting stuff and things
The square sliding window approach does not provide a natural way of working with irregular objects. Using region labeling as an image representation allows dealing with irregular
and extended objects (buildings, bookshelf, road, ...). Extended stuff [1] may be a very
important source of contextual information for other objects.
(a) Examples from the dictionary of about 2000 patches and masks, Ux,y , Vx,y .
(b) Examples from the dictionary of 30 graphs, Wx,y,c .
t=1
f
+
t=2
F
+
S
... =
Truth
f
Output
f
t=0
(c) Example feedforward segmentation for screens.
Figure 3: Examples of patches from the dictionary and an example of the segmentation obtained
using boosting trained with patches from (a).
The weak learners we use for the local evidence potentials are based on the segmentation
fragments proposed in [2]. Specifically, we create a dictionary of about 2000 image patches
U , chosen at random (but overlapping each object), plus a corresponding set of binary (inclass/ out-of-class) image masks, V : see Fig. 3(a). At each round t, for each class c, and
for each dictionary entry, we construct the following weak learner, whose output is a binary
matrix of the same size as the image I:
v(I) = ((I ? U ) > ?) ? V > 0
(9)
where ? represents normalized cross-correlation and ? represents convolution. The intuition behind this is that I ? U will produce peaks at image locations that contain this
patch/template, and then convolving with V will superimpose the segmentation mask on
top of the peaks. As a function of the threshold ?, the feature will behave more as a template
detector (? ? 1) or as a texture descriptor (? << 1).
To be able to detect objects at multiple scales, we first downsample the image to scale ?,
compute v(I ? ?), and then upsample the result. The final weak learner does this for
multiple scales, ORs all the results together, and then takes a linear transformation.
f (I) = ? (?? [v(I ? ?) ? ?]) + ?
(10)
Fig. 3(c) shows an example of segmentation obtained by using boosting without context.
The weak learners we use for the compatibility functions have a similar form:
!
C
X
gc (b) = ?
bc? ? Wc? + ?
(11)
c? =1
where bc? is the image formed by the beliefs at all pixels for class c? . This convolution
corresponds to eq. 8 in which the node i is one pixel x, y of class c. The binary kernels
(graph fragments) W define, for each node x, y of object class c, all the nodes from which it
will receive messages. These kernels are chosen by sampling patches of various sizes from
the labeling of images from the training set. This allows generating complicated patterns
of connectivity that reflect the statistics of object co-occurrences in the training set. The
overall incoming message is given by adding the kernels obtained at each boosting round.
(This is the key difference from mutual boosting [4], where the incoming message is just
the output of a single weak learner; thus, in mutual boosting, previously learned inter-class
connections are only used once.) Although it would seem to take O(t) time to compute Gt ,
we can precompute a single equivalent kernel W ? , so at runtime the overall complexity is
still linear in the number of boosting rounds, O(T ).
!
C
t
C
X
X
X def X
t
n
n
b c? ?
Gx,y,c =
bc? ? Wc?? + ? ?
? Wc? +
?n =
c? =1
n=1
n
c? =1
car car
building car
road car
F
Road
b=?(F+G)
Car
x
car building building building road building
Building
car road
building road
a) Incoming messages
to a car node.
b) Compatibilities (W?).
t=1
t=2
t=4
G
road road
y
t=20
c) A car out of context
(outside 3rd floor windows)
is less of a car.
t=40
Final labeling
b(car)
S(all)
d) Evolution of the beliefs for the car nodes (b) and labeling (S) for road, building, car.
Figure 4: Street scene. The BRF is trained to detect cars, buildings and the road.
In Fig. 4(a-b), we show the structures of the graph and the weights W ? defined by GT for
a BRF trained to detect cars, buildings and roads in street scenes.
3.2
Learning and inference
For training we used a labeled dataset of office and street scenes with about 100 images in
each set. During the training, in the first 5 rounds we only update the local potentials, to
allow local evidence to accrue. After the 5th iteration we start updating also the compatibility functions. At each round, we update only the local potential and compatibility function
associated with a single object class that reduces the most the multiclass cost. This allows
objects that need many features to have more complicated local potentials.
The algorithm learns to first detect easy (and large) objects, since these reduce the error of
all classes the fastest. The easy-to-detect objects can then pass information to the harder
ones. For instance, in office scenes, the system first detects screens, then keyboards, and
finally computer mice. Fig. 5 illustrates this behavior on the test set. A similar behavior is
obtained for the car detector (Fig. 4(d)). The detection of building and road provides strong
constraints for the locations of the car.
3.3
Cascade of classifiers with BRFs
The BRF can be turned into a cascade [12] by thresholding the beliefs. Computations
can then be reduced by doing the convolutions (required for computing f and g) only in
pixels that are still candidates for the presence of the target. At each round we update a
t
binary rejection mask for each object class, Rx,y,c
, by thresholding the beliefs at round t:
t
t?1
t
t
Rx,y,c = Rx,y,c ?(bx,y,c > ?c ). A pixel in the rejection mask is set to zero when we can
decide that the object is not present (when btx,y,c is below the threshold ?ct ? 0), and it is set
to 1 when more processing is required. The threshold ?ct is chosen so that the percentage
of missed detections is below a predefined level (we use 1%). Similarity we can define a
detection mask that will indicate pixels in which we decide the object is present. The mask
is then used for computing the features v(I) and messages G by applying the convolutions
only on the pixels not yet classified. We can denote those operators as ?R and ?R . This
Input image
screen
mouse
Ground truth
Output labeling
keyboard
t=5
b (screen)
t=10
t=15
b (screen)
b (screen)
t=25
b (screen)
t=50
b (screen)
F
G
b (keyboard)
b (keyboard)
b (keyboard)
b (keyboard)
b (keyboard)
b (mouse)
b (mouse)
b (mouse)
b (mouse)
b (mouse)
F
G
F
Area under ROC
G
1
Screen
BRF
0.5
t=0
Mouse
Keyboard
t=20
Boosting
Iteration (t)
t=50
Figure 5: Top. In this desk scene, it is easy to identify objects like the screen, keyboard and mouse,
even though the local information is sometimes insufficient. Middle: the evolution of the beliefs
(b and F and G) during detection for a test image. Bottom. The graph bellow shows the average
evolution of the area under the ROC for the three objects on 120 test images.
results in a more efficient classifier with only a slight decrease of performance. In Fig. 6 we
compare the reduction of the search space when implementing a cascade using independent
boosting (which reduces to Viola and Jones [12]), and when using BRF?s. We see that for
objects for which context is the main source of information, like the mouse, the reduction
in search space is much more dramatic using BRFs than using boosting alone.
4
Conclusion
The proposed BRF algorithm combines boosting and CRF?s, providing an algorithm that
is easy for both training and inference. We have demonstrated object detection in cluttered
scenes by exploiting contextual relationships between objects. The BRF algorithm is computationally efficient and provides a natural extension of the cascade of classifiers by integrating evidence from other objects in order to quickly reject certain image regions. The
BRF?s densely connected graphs, which efficiently collect information over large image
regions, provide an alternative framework to nearest-neighbor grids for vision problems.
Acknowledgments
This work was sponsored in part by the Nippon Telegraph and Telephone Corporation as
part of the NTT/MIT Collaboration Agreement, by BAE systems, and by DARPA contract
DABT63-99-1-0012.
Size of search space
Detection rate
Screen
Keyboard
Mouse
100%
Boosting
Boosting
Boosting
50%
BRF
BRF
BRF
Round
0%
0
1
10
20
30
40
Round
0
BRF
0.8
10
20
30
Round
0
20
30
40
Boosting
Boosting
0.5
10
BRF
BRF
Boosting
40
0.4
0.2
0
0
0.5
False alarm rate
1 0
0.5
False alarm rate
1 0
0.5
False alarm rate
1
Figure 6: Contextual information reduces the search space in the framework of a cascade and improves performances. The search space is defined as the percentage of pixels that require further
processing before a decision can be reached at each round. BRF?s provide better performance and
requires fewer computations. The graphs (search space and ROCs) correspond to the average results
on a test set of 120 images.
References
[1] E. H. Adelson. On seeing stuff: the perception of materials by humans and machines. In Proc.
SPIE, volume 4299, pages 1?12, 2001.
[2] E. Borenstein and S. Ullman. Class-specific, top-down segmentation. In Proc. European Conf.
on Computer Vision, 2002.
[3] T. Dietterich, A. Ashenfelter, and Y. Bulatov. Training conditional random fields via gradient
tree boosting. In Intl. Conf. on Machine Learning, 2004.
[4] M. Fink and P. Perona. Mutual boosting for contextual influence. In Advances in Neural Info.
Proc. Systems, 2003.
[5] J. Friedman, T. Hastie, and R. Tibshirani. Additive logistic regression: a statistical view of
boosting. Annals of statistics, 28(2):337?374, 2000.
[6] Xuming He, Richard Zemel, and Miguel Carreira-Perpinan. Multiscale conditional random
fields for image labelling. In Proc. IEEE Conf. Computer Vision and Pattern Recognition,
2004.
[7] S. Kumar and M. Hebert. Discriminative random fields: A discriminative framework for contextual interaction in classification. In IEEE Conf. on Computer Vision and Pattern Recognition,
2003.
[8] J. Lafferty, A. McCallum, and F. Pereira. Conditional random fields: Probabilistic models for
segmenting and labeling sequence data. In Intl. Conf. on Machine Learning, 2001.
[9] K. Murphy, A. Torralba, and W. Freeman. Using the forest to see the trees: a graphical model
relating features, objects and scenes. In Advances in Neural Info. Proc. Systems, 2003.
[10] R. Schapire. The boosting approach to machine learning: An overview. In MSRI Workshop on
Nonlinear Estimation and Classification, 2001.
[11] B. Taskar, C. Guestrin, and D. Koller. Max-margin markov networks. In Advances in Neural
Info. Proc. Systems, 2003.
[12] P. Viola and M. Jones. Robust real-time object detection. Intl. J. Computer Vision, 57(2):137?
154, 2004.
| 2663 |@word middle:1 seek:1 git:4 dramatic:1 harder:2 reduction:2 fragment:5 bc:4 current:1 contextual:9 si:22 yet:1 additive:4 wx:1 informative:3 sponsored:1 update:11 alone:1 selected:1 fewer:1 mccallum:1 detecting:2 boosting:28 node:15 location:3 gx:1 provides:2 become:1 combine:3 introduce:1 pairwise:1 inter:1 mask:7 behavior:2 examine:1 freeman:2 detects:1 bulatov:1 window:3 what:1 transformation:1 corporation:1 pseudo:2 every:1 stuff:6 ti:4 fink:2 runtime:2 classifier:4 murphyk:1 segmenting:1 before:1 local:19 brf:18 approximately:1 plus:1 collect:1 co:1 fastest:1 range:1 bi:3 acknowledgment:1 implement:1 procedure:1 area:2 cascade:6 reject:1 induce:1 road:13 seeing:2 integrating:1 operator:1 context:4 impossible:1 influence:2 applying:1 equivalent:1 map:1 demonstrated:1 crfs:3 l:2 cluttered:1 simplicity:1 searching:1 annals:1 pt:3 target:1 us:1 nippon:1 agreement:1 element:1 recognition:3 updating:1 labeled:3 v6t:1 bottom:2 taskar:1 region:6 connected:5 decrease:1 intuition:1 complexity:1 trained:3 solving:1 segment:1 learner:12 basis:1 joint:1 darpa:1 various:1 describe:2 zemel:1 labeling:9 kevin:1 neighborhood:1 outside:1 quite:1 encoded:1 whose:1 ability:1 statistic:2 gi:11 ip:1 final:2 blob:1 sequence:1 propose:4 interaction:1 product:1 coming:1 turned:1 exploiting:1 intl:3 produce:1 generating:1 object:38 help:1 miguel:1 ij:2 nearest:2 eq:4 strong:2 c:2 implies:1 indicate:1 correct:1 human:2 vx:1 material:1 implementing:1 require:1 extension:1 hold:1 ground:1 great:1 viterbi:1 dictionary:7 torralba:3 estimation:1 proc:6 label:3 create:1 weighted:5 mit:11 rather:1 boosted:4 office:3 derived:1 notational:1 likelihood:2 detect:8 inference:5 mrfs:1 downsample:1 bt:8 perona:2 hidden:1 koller:1 interested:1 pixel:12 compatibility:11 arg:2 classification:3 overall:2 spatial:1 initialize:2 mutual:3 marginal:1 field:13 construct:1 once:1 sampling:1 represents:3 jones:2 adelson:1 minimized:1 simplify:1 richard:1 densely:5 individual:1 relabeling:1 murphy:3 william:1 friedman:1 detection:14 interest:1 message:12 behind:2 predefined:1 tree:2 taylor:1 accrue:1 guidance:1 instance:1 cost:4 introducing:1 subset:2 entry:1 recognizing:2 successful:1 dependency:2 peak:2 csail:2 contract:1 probabilistic:1 telegraph:1 decoding:1 pool:1 together:1 mouse:13 quickly:1 connectivity:1 reflect:2 opposed:1 choose:1 conf:5 convolving:1 bx:1 ullman:1 account:1 potential:14 stump:3 piece:1 performed:1 view:1 doing:1 reached:1 start:1 complicated:2 minimize:2 square:3 ni:9 accuracy:1 convolutional:1 descriptor:1 formed:1 efficiently:1 correspond:1 identify:1 weak:14 rx:3 ah:1 classified:1 detector:3 explain:1 bni:1 mi:1 associated:1 spie:1 dataset:1 car:18 improves:1 segmentation:10 back:1 though:1 just:3 correlation:3 hand:1 working:1 nonlinear:1 multiscale:1 propagation:2 overlapping:1 logistic:1 stagewise:2 gti:7 building:13 dietterich:2 requiring:1 true:1 normalized:1 contain:1 evolution:3 hence:5 spatially:2 round:13 during:2 crf:4 complete:1 image:30 fi:17 sigmoid:2 overview:1 volume:1 assembling:1 he:2 slight:1 marginals:1 relating:1 cambridge:2 rd:1 grid:2 z4:1 nonlinearity:1 access:1 similarity:1 bti:12 gt:7 add:1 optimizing:1 keyboard:12 certain:1 brfs:7 success:1 binary:4 yi:4 exploited:1 guestrin:1 floor:1 maximize:1 sliding:2 multiple:3 infer:1 reduces:5 ntt:1 faster:1 cross:1 long:3 regression:4 btx:1 vision:7 iteration:10 kernel:4 sometimes:1 irregular:2 receive:1 addition:1 want:1 source:2 sends:2 borenstein:1 thing:3 lafferty:1 seem:1 presence:1 feedforward:1 easy:4 fit:15 architecture:2 topology:1 hastie:1 reduce:1 simplifies:2 idea:1 multiclass:3 fragile:1 passing:1 antonio:1 desk:1 reduced:1 schapire:1 percentage:2 msri:1 singleclass:1 per:1 yy:2 tibshirani:1 write:1 dropping:1 key:2 threshold:3 graph:19 bae:1 sum:1 run:1 almost:1 reasonable:1 decide:2 patch:7 missed:1 decision:1 capturing:1 bound:1 def:1 ct:2 followed:1 simplification:1 constraint:1 bp:1 scene:9 btk:4 wc:3 speed:1 min:2 kumar:1 combination:1 precompute:1 across:1 wi:3 s1:2 restricted:1 invariant:1 taken:1 computationally:1 equation:1 previously:1 discus:1 fail:1 wrt:1 efi:1 apply:1 indirectly:1 occurrence:1 alternative:1 top:4 graphical:1 exploit:2 build:2 approximating:1 dependence:1 traditional:1 gin:4 gradient:1 distance:1 street:4 jit:3 code:2 relationship:2 insufficient:1 providing:1 info:3 negative:1 btm:6 boltzmann:1 convolution:5 markov:1 behave:1 defining:1 extended:3 communication:1 viola:2 gc:1 superimpose:1 bk:2 pair:1 required:2 connection:8 learned:2 able:1 below:3 pattern:3 xm:3 perception:1 max:1 belief:18 natural:2 rely:1 improve:1 library:1 vancouver:1 billf:1 loss:1 xuming:1 thresholding:2 bank:1 collaboration:1 repeat:1 hebert:1 allow:2 neighbor:5 template:2 ashenfelter:1 bm:4 sj:1 dealing:1 global:1 sequentially:1 incoming:4 symmetrical:1 assumed:2 xi:11 discriminative:2 search:6 sk:15 learn:6 robust:1 forest:1 expansion:1 european:1 constructing:1 did:3 dense:2 main:1 whole:1 logitboost:1 s2:2 alarm:3 bookshelf:1 fig:9 roc:3 screen:11 fashion:1 pereira:1 candidate:1 perpinan:1 learns:1 down:2 specific:1 jt:1 evidence:10 workshop:1 false:3 adding:1 texture:1 labelling:1 illustrates:1 margin:1 rejection:2 likely:1 ux:1 upsample:1 ubc:2 truth:2 corresponds:1 ma:2 conditional:6 goal:3 change:1 carreira:1 typical:1 determined:1 specifically:1 telephone:1 bellow:1 pas:1 support:1 people:1 correlated:1 |
1,832 | 2,664 | Dynamic Bayesian Networks for
Brain-Computer Interfaces
Pradeep Shenoy
Department of Computer Science
University of Washington
Seattle, WA 98195
[email protected]
Rajesh P. N. Rao
Department of Computer Science
University of Washington
Seattle, WA 98195
[email protected]
Abstract
We describe an approach to building brain-computer interfaces (BCI)
based on graphical models for probabilistic inference and learning. We
show how a dynamic Bayesian network (DBN) can be used to infer
probability distributions over brain- and body-states during planning and
execution of actions. The DBN is learned directly from observed data
and allows measured signals such as EEG and EMG to be interpreted in
terms of internal states such as intent to move, preparatory activity, and
movement execution. Unlike traditional classification-based approaches
to BCI, the proposed approach (1) allows continuous tracking and prediction of internal states over time, and (2) generates control signals based
on an entire probability distribution over states rather than binary yes/no
decisions. We present preliminary results of brain- and body-state estimation using simultaneous EEG and EMG signals recorded during a
self-paced left/right hand movement task.
1
Introduction
The problem of building a brain-computer interface (BCI) has received considerable attention in recent years. Several researchers have demonstrated the feasibility of using EEG
signals as a non-invasive medium for building human BCIs [1, 2, 3, 4, 5] (see also [6] and
articles in the same issue). A central theme in much of this research is the postulation of
a discrete brain state that the user maintains while performing one of a set of physical or
imagined actions. The goal is to decode the hidden brain state from the observable EEG
signal, and to use the decoded state to control a robot or a cursor on a computer screen.
Most previous approaches to BCI (e.g., [1, 2, 4]) have utilized classification methods applied to time slices of EEG data to discriminate between a small set of brain states (e.g., left
versus right hand movement). These methods typically involve various forms of preprocessing (such as band-pass filtering or temporal smoothing) as well as feature extraction on
time slices known to contain one of the chosen set of brain states. The output of the classifier is typically a yes/no decision regarding class membership. A significant drawback of
such an approach is the need to have a ?point of reference? for the EEG data, i.e., a synchronization point in time where the behavior of interest was performed. Also, classifier-based
approaches typically do not model the uncertainty in their class estimates. As a result, it
is difficult to have a continuous estimate of the brain state and to associate an uncertainty
with the current estimate.
In this paper, we propose a new framework for BCI based on probabilistic graphical models [7] that overcomes some of the limitations of classification-based approaches to BCI.
We model the dynamics of hidden brain- and body-states using a Dynamic Bayesian Network (DBN) that is learned directly from EEG and EMG data. We show how a DBN can be
used to infer probability distributions over hidden state variables, where the state variables
correspond to brain states useful for BCI (such as ?Intention to move left hand?, ?Left hand
in motion?, etc). Using a DBN gives us several advantages in addition to providing a continuous probabilistic estimate of brain state. First, it allows us to explicitly model the hidden
causal structure and dependencies between different brain states. Second, it facilitates the
integration of information from multiple modalities such as EEG and EMG signals, allowing, for example, EEG-derived estimates to be bootstrapped from EMG-derived estimates.
In addition, learning a dynamic graphical model for time-varying data such as EEG allows
other useful operations such as prediction, filling in of missing data, and smoothing of state
estimates using information from future data points. These capabilities are difficult to obtain while working exclusively in the frequency domain or using whole slices of the data
(or its features) for training classifiers. We illustrate our approach in a simple Left versus
Right hand movement task and present preliminary results showing supervised learning and
Bayesian inference of hidden state for a dataset containing simultaneous EEG and EMG
recordings.
2
The DBN Framework
We study the problem of modeling spontaneous movement of the left/right arm using EEG
and EMG signals. It is well known that EEG signals show a slow potential drift prior to
spontaneous motor activity. This potential drift, known as the Bereitschaftspotential (BP,
see [8] for an excellent survey), shows variation in distribution over scalp with respect to
the body part being moved. In particular, the BP related to movement of left versus right
arm shows a strong lateral asymmetry. This allows one to not only estimate the intent to
move prior to actual movement, but also distinguish between left and right movements.
Previous approaches [1, 2] have utilized BP signals in classification-based BCI protocols
based on synchronization cues that identify points of movement onset. In our case, the
challenge was to model the structure of BPs and related movement signals using the states
of the DBN, and to recognize actions without explicit synchronization cues.
Figure 1 shows the complete DBN (referred to as Nf ull in this paper) used to model the leftright hand movement task. The hidden state Bt in Figure 1(a) tracks the higher-level brain
state over time and generates the hidden EEG and EMG states Et and Mt respectively.
These hidden states in turn generate the observed EEG and EMG signals. The dashed
arrows indicate that the hidden states make transitions over time. As shown in Figure 1(b),
the state Bt is intended to model the high-level intention of the subject. The figure shows
both the values Bt can take as well the constraints on the transition between values. The
actual probabilities of the allowed transitions are learned from data.
The hidden states Et and Mt are intended to model the temporal structure of the EEG and
EMG signals, which are generated using a mixture of Gaussians conditioned on E t and
Mt respectively. In the same way as the values of Bt are customized for our particular
experiment, we would like the state transitions of Et and Mt to also reflect their respective
constraints. This is important since it allows us to independently learn the simpler DBN
Nemg consisting of only the node Mt and the observed EMG signal. Similarly, we can
also independently learn the model Neeg consisting of the node Et and the observed EEG
signal. We use the models shown in Figure 2 for allowed transitions of the states M t and Et
respectively. In particular, Figure 2(a) indicates that the EMG state can transition along one
transition
Bt
Bt+1
Et
transition
Et+1
Left
Movt
Post
Movt
Rt
Intent
Rt
Movt
Post
Movt
Rest
Mt
EEGt
Left
Intent
Mt+1
EEGt+1
EMGt
EMGt+1
(b) Allowed Brain States Bt
(a) The Complete Network
Figure 1: Dynamic graphical model for modeling brain and body processes in a selfpaced movement task: (a) At each time instant t, the brain state Bt generates the EEG
and EMG internal states Et and Mt respectively, which in turn generate the observed EEG
and EMG. The dotted arrows represent transitions to a state at the next time step. (b) The
transition graph for the brain state Bt . The probability of each allowed transition is learned
from input data.
of three chains of states (labeled (1), (2), and (3)), representing the rest state, a left-hand
action and a right-hand action respectively. In each chain, the state Mt in each time step
either retains its old value with a given probability (self-pointing arrow) or transitions to the
next state value in that particular chain. The transition graph of Figure 2(b) shows similar
constraints on the EEG, except that the left and right action chains are further partitioned
into intent, action, and post-action subgroups of states, since each of these components are
discernible from the BP in EEG (but not from EMG) signals.
(chain of states)
e1
(chain of states)
mp+1
mq
(2)
(1)
m0
mq+1
(chain of states)
(a) Transition Constraints on Mt
LM
LPM
RM
RPM
(2)
(1)
e0
(3)
(3)
mp
(chain of states)
(chain of states)
m1
LI
ek
mr
RI
(chain of states)
(b) Transition Constraints onEt
Figure 2: Constrained transition graphs for the hidden EMG and EEG states E t and
Mt respectively. (a) The EMG state transitions between its values mi are constrained to
be in one of three chains: the chains model (1) rest, (2) left arm movement, and (3) right
arm movement. (b) In the EEG state transition graph, the left and right movement chains
are further divided into state values encoding intent (LI/RI), movement (LM/RM), and post
movement (LPM/RPM).
3
3.1
Experiments and Results
Data Collection and Processing
The task: The subject pressed two distinct keys on a keyboard with the left hand or right
hand at random at a self-initiated pace. We recorded 8 EEG channels around the motor area
of cortex (C3, Cz, C4, FC1, FC2, CP1, CP2, Pz) using averaged ear electrodes as reference,
and 2 differential pairs of EMG (one on each arm). Data was recorded at 2048Hz for a
period of 20 minutes, with the movements being separated by approximately 3-4s.
Average for Left Hand
Average for Right Hand
4
2
2
0
0
?2
?2
?4
C3
C4
?6
?4
C3
C4
?6
?0.5
0
Time
0.5
1
?0.5
0
Time
0.5
1
Figure 3: Movement-related potential drift recorded during the hand-movement task:
The two plots show the EEG signals averaged over all trials from the motor-related channels
C3 and C4 for left (left panel) and right hand movement (right panel). The averages indicate
the onset and laterality of upcoming movements.
Processing: The EEG channels were bandpass-filtered 0.5Hz-5Hz, before being downsampled and smoothed at 128Hz. The EMG channels were converted to RMS values computed
over windows for an effective sampling rate of 128Hz.
Data Analysis: The recorded data were first analyzed in the traditional manner by averaging across all trials. Figure 3 shows the average of EEG channels C3 and C4 for left
and right hand movement actions respectively. As can be seen, the averages for both channels are different for the two classes. Furthermore, there is a slow potential drift preceding
the action and a return to the baseline potential after the action is performed. Previous
researchers [1] have classified EEG data over a window leading up to the instant of action
with high accuracy (over 90%) into left or right movement classes. Thus, there appears to
be a reliable amount of information in the EEG signal for at least discriminating between
left versus right movements.
Data Evaluation using SVMs: To obtain a baseline and to evaluate the quality of our
recorded data, we tested the performance of linear support vector machines (SVMs) on
classifying our EEG data into left and right movement classes. The choice of linear SVMs
was motivated by their successful use on similar problems by other researchers [1]. Time
slices of 0.5 seconds before each movement were concatenated from all EEG channels and
used for classification. We performed hyper-parameter selection using leave-one-out crossvalidation on 15 minutes of data and obtained an error of 15% on the remaining 5 minutes
of data. Such an error rate is comparable to those obtained in previous studies on similar
tasks, suggesting that the recorded data contains sufficient movement-related information
to be tested in experiments involving DBNs.
Learning the parameters of the DBN: We used the Graphical Models Toolkit
(GMTK) [9] for learning the parameters of our DBN. GMTK provides support for expressing constraints on state transitions (as described in Section 2). It learns the constrained
conditional probability tables and the parameters for the mixture of Gaussians using the
expectation-maximization (EM) algorithm.
We constructed a supervisory signal from the recorded key-presses as follows: A period of
100ms around each keystroke was labeled ?motor action? for the appropriate hand. This
signal was used to train the network Nemg in a supervised manner. To generate a supervisory signal for the network Neeg , or the full combined network Nf ull (Figure 1), we
added prefixes and postfixes of 150ms each to each action in this signal, and labeled them
?preparatory? and ?post-movement? activity respectively. These time-periods were chosen
by examining the average EEG and EMG activity over all actions. Thus, we can use partial
(EEG only) or full evidence in the inference step to obtain probability distributions over
brain state. The following sections describe our learning procedure and inference results in
greater detail.
3.2
Learning and Inference with EMG
Our first step is to learn the simpler model Nemg that has only the hidden Mt state and the
observed EMG signal. This is to test inference using the EMG signal alone. The parameters
of this DBN were learned in a supervised manner.
We used 15 minutes of EMG data to train our simplified model, and then tested it on the
remaining 5 minutes of data. The model was tested using Viterbi decoding (a single pass
of max-product inference over the network). In other words, the maximum a posteriori
(MAP) sequence of values for hidden states was computed. Figure 4 shows a 100s slice of
data containing 2 channels of EMG, and the predicted hidden EMG state Mt . The states 0,
1 and 2 correspond to ?no action?, left, and right actions respectively. In the shown figure,
the state Mt successfully captures not only all the obvious arm movements but also the
actions that are obscured by noise.
3.3
Learning the EEG Model
We used the supervisory signal described earlier to learn the corresponding EEG model
Neeg . Note that the brain-state can be inferred from the hidden EEG state Et directly, since
the state space is appropriately partitioned as shown in Figure 2(b).
Figure 5 shows the result of inference on the learned model Neeg using only the EEG
signals as evidence. The figure shows a subset of the EEG channels (C3,Cz,C4), the supervisory signal, and the predicted brain state Bt (the MAP estimate). The figure shows that
many of the instances of action (but not all) are correctly identified by the model.
Our model gives us at each time instant a MAP-estimated state sequence that best describes
the past, and the probability associated with that state sequence. This gives us, at each time
instant, a measure of how likely each brain state Bt is, with reference to the others. For
convenience, we can use the probability associated with the REST state (see Figure 1) as
reference. Figure 6 shows a graphical illustration of this instantaneous time estimate. The
plotted graphs are, in order, the supervisory signal (i.e., the ?ground truth value?) and the
instantaneous measures of likelihood of intention/movement/post-movement states for the
left and right hand respectively. For convenience, we represent the likelihood ratio of each
state?s MAP probability estimate to that of the rest state, and use a logarithmic scale. We
see that the true hand movements are correctly inferred in a surprisingly large number of
cases (log likelihood ratio crosses 0). Furthermore, the actual likelihood values convey a
measure of the uncertainty in the inference, a property that would be of great value for
critical BCI applications such as controlling a robotic wheelchair.
In summary, our graphical models Nemg and Neeg have shown promising results in correctly identifying movement onset from EMG and EEG signals respectively. Ongoing work
is focused on improving accuracy by using features extracted from EEG, and inference using both EEG and EMG in Nf ull (the full model).
10
20
30
40
50
60
70
80
90
100
10
20
30
40
50
60
70
80
90
100
10
20
30
40
50
60
Time (seconds)
70
80
90
100
30
20
10
Left Hand EMG
Predicted State
Right Hand EMG
14
12
10
8
6
4
2
2
1
0
Figure 4: Bayesian Inference of Movement using EMG: The figure shows 100 seconds of
EMG data from two channels along with the MAP state sequence predicted by our trained
EMG model. The states 0,1,2 correspond to ?no action?, left, and right actions respectively.
Our model correctly identifies the obscured spikes in the noisy right EMG channel
4
Discussion and Conclusion
We have shown that dynamic Bayesian networks (DBNs) can be used to model the transitions between brain- and muscle-states as a subject performs a motor task. In particular,
a two-level hierarchical network was proposed for simultaneously estimating higher-level
brain state and lower-level EEG and EMG states in a left/right hand movement task. The
results demonstrate that for a self-paced movement task, hidden brain states useful for BCI
such as intention to move the left or right hand can be decoded from a DBN learned directly
from EEG and EMG data.
Previous work on BCIs can be grouped into two broad classes: self-regulatory BCIs and
BCIs based on detecting brain state. Self-regulatory BCIs rely on training the user to regulate certain features of the EEG, such as cortical positivity [10], or oscillatory activity (the
? rhythm, see [5]), in order to control, for example, a cursor on a display. The approach
presented in this paper falls in the second class of BCIs, those based on detecting brain
states [1, 2, 3, 4]. However, rather than employing classification methods, we use probabilistic graphical models for inferring brain state and learning the transition probabilities
between brain states.
Successfully learning a dynamic graphical model as suggested in this paper offers several
advantages over traditional classification-based schemes for BCI. It allows one to explicitly model the hidden causal structure and dependencies between different brain states. It
provides a probabilistic framework for integrating information from multiple modalities
2
0
?2
C3
0
1000
2000
3000
4000
5000
6000
7000
8000
Cz ?20
1000
2000
3000
4000
5000
6000
7000
8000
0
1000
2000
3000
4000
5000
6000
7000
8000
0
True Brain
State
1000
2000
3000
4000
5000
6000
7000
8000
1000
2000
3000
4000
5000
6000
7000
8000
2
0
2
0
C4
?2
6
4
2
0
6
4
2
0
0
Predicted Brain State
Figure 5: Bayesian Inference of Brain State using EEG: The figure shows 1 minute of
EEG data (at 128Hz) for the channels C3, Cz, C4, along with the ?true? brain state and
the brain state inferred using our DBN model with only EEG evidence. State 0 is the rest
state, states 1 through 3 represent left hand movement, and 4 through 6 represent right hand
movement (see Figure 1(b)).
such as EEG and EMG signals, allowing, for example, EEG-derived estimates to be bootstrapped from EMG-derived estimates. A dynamic graphical model for time-varying data
such as EEG also allows prediction, filling in of missing data, and smoothing of state estimates using information from future data points, properties not easily achieved in methods
that work exclusively in the frequency domain or use data slices for training classifiers. Our
current efforts are focused on investigating methods for learning dynamic graphical models
for motor tasks of varying complexity and using these models to build robust, probabilistic
BCI systems.
References
[1] B. Blankertz, G. Curio, and K.R. Mueller. Classifying single trial EEG: Towards
brain computer interfacing. In Advances in Neural Information Processing Systems
12, 2001.
[2] G. Dornhege, B. Blankertz, G. Curio, and K.-R. Mueller. Combining features for
BCI. In Advances in Neural Information Processing Systems 15, 2003.
[3] J. D. Bayliss and D. H. Ballard. Recognizing evoked potentials in a virtual environment. In Advances in Neural Information Processing Systems 12, 2000.
[4] P. Meinicke, M. Kaper, F. Hoppe, M. Heumann, and H. Ritter. Improving transfer
rates in brain computer interfacing: a case study. In Advances in Neural Information
Processing Systems 15, 2003.
[5] J.R. Wolpaw, D.J. McFarland, and T.M. Vaughan. Brain-computer interfaces for communication and control. IEEE Trans Rehab Engg, pages 222?226, 2000.
6
4
2
0
0
500
1000
1500
2000
2500
3000
3500
4000
Supervisory Signal
20
0
?20
?40
?60
?80
0
500
1000
1500
2000
2500
3000
3500
Left action Log(P(state)/P(rest))
zero
pre
movt
post
4000
20
0
?20
?40
?60
?80
0
500
1000
1500
2000
2500
3000
3500
zero
pre
movt
post
4000
Right action Log(P(state)/P(rest))
Figure 6: Probabilistic Estimation of Brain State: The figure shows the supervisory
signal, along with a probabilistic measure of the current state for left and right actions
respectively. The measure shown is the log ratio of the instantaneous MAP estimate for the
relevant state and the estimate for the rest state.
[6] J. R. Wolpaw et al. Brain-computer interface technology: a review of the first international meeting. IEEE Trans Rehab Engg, 8:164?173, 2000.
[7] R. E. Neapolitan. Learning Bayesian Networks. Prentice Hall, NJ, 2004.
[8] M. Jahanshahi and M. Hallet. The Bereitschaftspotential: movement related cortical
potentials. Kluwer Academic, New York, 2002.
[9] J. Bilmes and G. Zweig. The graphical models toolkit: An open source software system for speech and time-series processing. In IEEE Intl. Conf. on Acoustics, Speech,
and Signal Processing, Orlando FL, 2002.
[10] N. Birbaumer, N. Ghanayim, T. Hinterberger, I. Iverson, B. Kotchubey, A. Kiibler,
J. Perelmouter, E. Taub, and H. Flor. A spelling device for the paralyzed. In Nature,
398: 297-298, 1999.
| 2664 |@word trial:3 meinicke:1 open:1 pressed:1 cp2:1 contains:1 exclusively:2 series:1 bootstrapped:2 prefix:1 past:1 current:3 engg:2 discernible:1 motor:6 plot:1 alone:1 cue:2 device:1 filtered:1 provides:2 detecting:2 node:2 simpler:2 along:4 constructed:1 iverson:1 differential:1 manner:3 behavior:1 preparatory:2 planning:1 brain:41 actual:3 window:2 estimating:1 panel:2 medium:1 interpreted:1 nj:1 dornhege:1 temporal:2 nf:3 ull:3 classifier:4 rm:2 control:4 shenoy:1 before:2 encoding:1 initiated:1 approximately:1 evoked:1 averaged:2 wolpaw:2 procedure:1 area:1 intention:4 word:1 integrating:1 downsampled:1 pre:2 convenience:2 selection:1 prentice:1 vaughan:1 map:6 demonstrated:1 missing:2 attention:1 independently:2 survey:1 focused:2 identifying:1 mq:2 variation:1 spontaneous:2 dbns:2 controlling:1 user:2 decode:1 associate:1 utilized:2 labeled:3 observed:6 postulation:1 capture:1 movement:40 environment:1 complexity:1 onet:1 dynamic:10 trained:1 easily:1 various:1 train:2 separated:1 distinct:1 describe:2 effective:1 hyper:1 gmtk:2 bci:13 bereitschaftspotential:2 noisy:1 advantage:2 sequence:4 propose:1 product:1 rehab:2 relevant:1 combining:1 moved:1 crossvalidation:1 seattle:2 electrode:1 asymmetry:1 intl:1 leave:1 illustrate:1 measured:1 heumann:1 received:1 strong:1 c:2 predicted:5 indicate:2 drawback:1 human:1 virtual:1 orlando:1 preliminary:2 around:2 hall:1 ground:1 great:1 viterbi:1 lm:2 pointing:1 m0:1 estimation:2 grouped:1 successfully:2 interfacing:2 rather:2 varying:3 derived:4 indicates:1 likelihood:4 baseline:2 posteriori:1 inference:12 mueller:2 membership:1 entire:1 typically:3 bt:11 hidden:17 issue:1 classification:7 smoothing:3 integration:1 constrained:3 extraction:1 washington:4 sampling:1 broad:1 filling:2 future:2 others:1 simultaneously:1 recognize:1 intended:2 consisting:2 interest:1 evaluation:1 mixture:2 analyzed:1 pradeep:1 chain:13 rajesh:1 partial:1 respective:1 old:1 plotted:1 causal:2 e0:1 obscured:2 instance:1 modeling:2 earlier:1 rao:2 retains:1 maximization:1 subset:1 recognizing:1 successful:1 examining:1 dependency:2 emg:38 perelmouter:1 combined:1 international:1 discriminating:1 ghanayim:1 probabilistic:8 ritter:1 decoding:1 central:1 recorded:8 reflect:1 containing:2 ear:1 positivity:1 hinterberger:1 conf:1 ek:1 leading:1 selfpaced:1 return:1 li:2 suggesting:1 potential:7 converted:1 explicitly:2 mp:2 onset:3 performed:3 maintains:1 capability:1 accuracy:2 correspond:3 identify:1 yes:2 bayesian:8 bilmes:1 researcher:3 classified:1 simultaneous:2 oscillatory:1 frequency:2 invasive:1 obvious:1 associated:2 mi:1 wheelchair:1 dataset:1 appears:1 higher:2 supervised:3 furthermore:2 hand:24 working:1 quality:1 bcis:6 supervisory:7 laterality:1 building:3 contain:1 true:3 during:3 self:6 rhythm:1 m:2 complete:2 demonstrate:1 performs:1 motion:1 interface:5 instantaneous:3 mt:14 physical:1 birbaumer:1 imagined:1 m1:1 pshenoy:1 kluwer:1 significant:1 expressing:1 taub:1 dbn:14 similarly:1 lpm:2 toolkit:2 robot:1 cortex:1 etc:1 recent:1 keyboard:1 certain:1 binary:1 meeting:1 muscle:1 seen:1 greater:1 preceding:1 mr:1 period:3 signal:32 dashed:1 paralyzed:1 multiple:2 full:3 infer:2 academic:1 cross:1 offer:1 zweig:1 divided:1 post:8 e1:1 feasibility:1 prediction:3 involving:1 hoppe:1 expectation:1 represent:4 cz:4 achieved:1 addition:2 source:1 modality:2 appropriately:1 flor:1 rest:9 unlike:1 recording:1 subject:3 hz:6 facilitates:1 identified:1 regarding:1 motivated:1 rms:1 effort:1 speech:2 york:1 action:24 useful:3 involve:1 amount:1 band:1 svms:3 generate:3 cp1:1 dotted:1 estimated:1 track:1 correctly:4 pace:1 discrete:1 key:2 graph:5 year:1 uncertainty:3 decision:2 rpm:2 comparable:1 fl:1 paced:2 distinguish:1 display:1 activity:5 scalp:1 constraint:6 bp:4 ri:2 software:1 generates:3 performing:1 department:2 across:1 describes:1 em:1 partitioned:2 turn:2 operation:1 gaussians:2 hierarchical:1 appropriate:1 regulate:1 remaining:2 graphical:12 instant:4 concatenated:1 build:1 upcoming:1 move:4 added:1 spike:1 rt:2 spelling:1 traditional:3 lateral:1 fc2:1 illustration:1 providing:1 ratio:3 difficult:2 intent:6 allowing:2 communication:1 postfix:1 smoothed:1 drift:4 inferred:3 pair:1 c3:8 c4:8 acoustic:1 learned:7 subgroup:1 trans:2 suggested:1 mcfarland:1 challenge:1 reliable:1 max:1 critical:1 rely:1 customized:1 arm:6 representing:1 scheme:1 blankertz:2 technology:1 fc1:1 identifies:1 prior:2 review:1 synchronization:3 limitation:1 filtering:1 versus:4 sufficient:1 article:1 classifying:2 summary:1 surprisingly:1 fall:1 slice:6 cortical:2 transition:21 collection:1 preprocessing:1 simplified:1 employing:1 observable:1 overcomes:1 robotic:1 investigating:1 continuous:3 regulatory:2 table:1 promising:1 ballard:1 nature:1 learn:4 channel:12 robust:1 transfer:1 eeg:51 improving:2 excellent:1 domain:2 protocol:1 arrow:3 whole:1 noise:1 allowed:4 convey:1 body:5 referred:1 bayliss:1 screen:1 slow:2 theme:1 decoded:2 explicit:1 bandpass:1 inferring:1 learns:1 minute:6 keystroke:1 showing:1 pz:1 evidence:3 curio:2 execution:2 conditioned:1 cursor:2 logarithmic:1 likely:1 tracking:1 truth:1 extracted:1 conditional:1 goal:1 towards:1 considerable:1 except:1 averaging:1 discriminate:1 pas:2 bps:1 internal:3 support:2 ongoing:1 evaluate:1 tested:4 |
1,833 | 2,665 | Hierarchical Eigensolver for Transition Matrices
in Spectral Methods
?
Chakra Chennubhotla? and Allan D. Jepson?
Department of Computational Biology, University of Pittsburgh
?
Department of Computer Science, University of Toronto
Abstract
We show how to build hierarchical, reduced-rank representation for large
stochastic matrices and use this representation to design an efficient algorithm for computing the largest eigenvalues, and the corresponding
eigenvectors. In particular, the eigen problem is first solved at the coarsest level of the representation. The approximate eigen solution is then
interpolated over successive levels of the hierarchy. A small number of
power iterations are employed at each stage to correct the eigen solution.
The typical speedups obtained by a Matlab implementation of our fast
eigensolver over a standard sparse matrix eigensolver [13] are at least a
factor of ten for large image sizes. The hierarchical representation has
proven to be effective in a min-cut based segmentation algorithm that we
proposed recently [8].
1 Spectral Methods
Graph-theoretic spectral methods have gained popularity in a variety of application domains: segmenting images [22]; embedding in low-dimensional spaces [4, 5, 8]; and clustering parallel scientific computation tasks [19]. Spectral methods enable the study of properties global to a dataset, using only local (pairwise) similarity or affinity measurements between the data points. The global properties that emerge are best understood in terms of a
random walk formulation on the graph. For example, the graph can be partitioned into clusters by analyzing the perturbations to the stationary distribution of a Markovian relaxation
process defined in terms of the affinity weights [17, 18, 24, 7]. The Markovian relaxation
process need never be explicitly carried out; instead, it can be analytically expressed using
the leading order eigenvectors, and eigenvalues, of the Markov transition matrix.
In this paper we consider the practical application of spectral methods to large datasets. In
particular, the eigen decomposition can be very expensive, on the order of O(n 3 ), where n
is the number of nodes in the graph. While it is possible to compute analytically the first
eigenvector (see ?3 below), the remaining subspace of vectors (necessary for say clustering)
has to be explicitly computed. A typical approach to dealing with this difficulty is to first
sparsify the links in the graph [22] and then apply an efficient eigensolver [13, 23, 3].
In comparison, we propose in this paper a specialized eigensolver suitable for large stochastic matrices with known stationary distributions. In particular, we exploit the spectral properties of the Markov transition matrix to generate hierarchical, successively lower-ranked
approximations to the full transition matrix. The eigen problem is solved directly at the
coarsest level of representation. The approximate eigen solution is then interpolated over
successive levels of the hierarchy, using a small number of power iterations to correct the
solution at each stage.
2 Previous Work
One approach to speeding up the eigen decomposition is to use the fact that the columns
of the affinity matrix are typically correlated. The idea then is to pick a small number
of representative columns to perform eigen decomposition via SVD. For example, in the
Nystrom approximation procedure, originally proposed for integral eigenvalue problems,
the idea is to randomly pick a small set of m columns; generate the corresponding affinity
matrix; solve the eigenproblem and finally extend the solution to the complete graph [9, 10].
The Nystrom method has also been recently applied in the kernel learning methods for
fast Gaussian process classification and regression [25]. Other sampling-based approaches
include the work reported in [1, 2, 11].
Our starting point is the transition matrix generated from affinity weights and we show
how building a representational hierarchy follows naturally from considering the stochastic matrix. A closely related work is the paper by Lin on reduced rank approximations of
transition matrices [14]. We differ in how we approximate the transition matrices, in particular our objective function is computationally less expensive to solve. In particular, one
of our goals in reducing transition matrices is to develop a fast, specialized eigen solver for
spectral clustering. Fast eigensolving is also the goal in ACE [12], where successive levels
in the hierarchy can potentially have negative affinities. A graph coarsening process for
clustering was also pursued in [21, 3].
3 Markov Chain Terminology
We first provide a brief overview of the Markov chain terminology here (for more details
see [17, 15, 6]). We consider an undirected graph G = (V, E) with vertices vi , for i =
{1, . . . , n}, and edges ei,j with non-negative weights ai,j . Here the weight ai,j represents
the affinity between vertices vi and vj . The affinities are represented by a non-negative,
symmetric n ? n matrix
ai,j as elements. The degree of a node j is
PnA having weights
Pn
defined to be: dj =
a
=
a
, where we define D = diag(d1 , . . . , dn ).
i,j
j,i
i=1
j=1
A Markov chain is defined using these affinities by setting a transition probability matrix
M = AD ?1 , where the columns of M each sum to 1. The transition probability matrix
defines the random walk of a particle on the graph G.
The random walk need never be explicitly carried out; instead, it can be analytically expressed using the leading order eigenvectors, and eigenvalues, of the Markov transition
matrix. Because the stochastic matrices need not be symmetric in general, a direct eigen
decomposition step is not preferred for reasons of instability. This problem is easily circumvented by considering a normalized affinity matrix: L = D ?1/2 AD?1/2 , which is related
to the stochastic matrix by a similarity transformation: L = D ?1/2 M D1/2 . Because L
is symmetric, it can be diagonalized: L = U ?U T , where U = [~u1 , ~u2 , ? ? ? , ~un ] is an
orthogonal set of eigenvectors and ? is a diagonal matrix of eigenvalues [?1 , ?2 , ? ? ? , ?n ]
sorted in decreasing order. The eigenvectors have unit length k~uk k = 1 and from the
form of A and D it can be shown that the eigenvalues ?i ? (?1, 1], with at least one
eigenvalue equal to one. Without loss of generality, we take ?1 = 1. Because L and M
are similar we can perform an eigen decomposition of the Markov transition matrix as:
M = D1/2 LD?1/2 = D1/2 U ? U T D?1/2 . Thus an eigenvector ~u of L corresponds to
an eigenvector D 1/2 ~u of M with the same eigenvalue ?.
The Markovian relaxation process after ? iterations, namely M ? , can be represented as:
M ? = D1/2 U ?? U T D?1/2 . Therefore, a particle undertaking a random walk with an
initial distribution p~ 0 acquires after ? steps a distribution p~ ? given by: p~ ? = M ? p~ 0 .
Assuming the graph is connected, as ? ? ?,
Pnthe Markov chain approaches a unique
stationary distribution given by ~? = diag(D)/ i=1 di , and thus, M ? = ~? 1T , where 1 is
a n-dim column vector of all ones. Observe that ~? is an eigenvector of M as it is easy to
show that M~? = ~? and the corresponding eigenvalue is 1. Next, we show how to generate
hierarchical, successively low-ranked approximations for the transition matrix M .
4 Building a Hierarchy of Transition Matrices
The goal is to generate a very fast approximation, while simultaneously achieving sufficient
f as some
accuracy. For notational ease, we think of M as a fine-scale representation and M
f further, we can generate
coarse-scale approximation to be derived here. By coarsening M
successive levels of the representation hierarchy. We use the stationary distribution ~? to
construct a corresponding coarse-scale stationary distribution ~?. As we just discussed a
critical property of the fine scale Markov matrix M is that it is similar to the symmetric
matrix L and we wish to preserve this property at every level of the representation hierarchy.
4.1 Deriving Coarse-Scale Stationary Distribution
We begin by expressing the stationary distribution ~? as a probabilistic mixture of latent
distributions. In matrix notation, we have
~? = K ~?,
(1)
where ~? is an unknown mixture coefficient vector of length m, K is an n ? P
m non-negative
n
kernel matrix whose columns are latent distributions that each sum to 1: i=1 Ki,j = 1
~
and m n. It is easy to derive a maximum likelihood approximation
of ? using an EM
~
type algorithm [16]. The main step is to find a stationary point ?, ? for the Lagrangian:
E??
n
X
i=1
?i ln
m
X
Ki,j ?j + ?
j=1
X
m
j=1
?j ? 1 .
(2)
An implicit step in this EM procedure is to compute the the ownership probability r i,j of
the j th kernel (or node) at the coarse scale for the ith node on the fine scale and is given by
?j Ki,j
.
ri,j = Pm
k=1 ?k Ki,k
(3)
The EM procedure allows for an update of both ~? and the latent distributions in the kernel
matrix K (see ?8.3.1 in [6]).
For initialization, ~? is taken to be uniform over the coarse-scale states. But in choosing
kernels K, we provide a good initialization for the EM procedure. Specifically, the Markov
matrix M is diffused using a small number of iterations to get M ? . The diffusion causes
random walks from neighboring nodes to be less distinguishable. This in turn helps us
select a small number of columns of M ? in a fast and greedy way to be the kernel matrix
K. We defer the exact details on kernel selection to a later section (?4.3).
4.2 Deriving the Coarse-Scale Transition Matrix
f, the coarse-scale transition matrix, we break it down into three steps.
In order to define M
First, the Markov chain propagation at the coarse scale can be defined as:
f~q k ,
~q k+1 = M
(4)
where ~q is the coarse scale probability distribution after k steps of the random walk.
Second, we expand ~q k into the fine scale using the kernels K resulting in a fine scale
probability distribution p~ k :
p~ k = K~q k .
(5)
k
Finally, we lift p~ k back into the coarse scale by using the ownership probability of the j th
kernel for the ith node on the fine grid:
qjk+1 =
n
X
i=1
ri,j pik
(6)
Substituting for Eqs.(3) and (5) in Eq. 6 gives
qjk+1 =
n
X
i=1
ri,j
m
X
Ki,t qtk =
t=1
n
X
i=1
? K
Pm j i,j
k=1 ?k Ki,k
X
m
Ki,t qtk .
(7)
t=1
We can write the preceding equation in a matrix form:
?1
K~q k .
~q k+1 = diag( ~? ) K T diag K ~?
f as:
Comparing this with Eq. 4, we can derive the transition matrix M
?1
f = diag( ~? ) K T diag K ~?
M
K.
(8)
(9)
f~?, so ~? is the stationary distribution for M
f. Following the
It is easy to see that ~? = M
~
f
definition of M , and its stationary distribution ?, we can generate a symmetric coarse scale
e given by
affinity matrix A
?1
e=M
fdiag(~?) = diag( ~? ) K T
diag K ~?
A
Kdiag(~?) ,
(10)
f from Eq. 9. The coarse-scale affinity matrix A
e is
where we substitute for the expression M
then normalized to get:
e=D
e ?1/2 A
eD
e ?1/2 ;
L
e = diag(de1 , de2 , ? ? ? , dem ),
D
(11)
e (see
where dej is the degree of node j in the coarse-scale graph represented by the matrix A
f is precisely similar to a
?3 for degree definition). Thus, the coarse scale Markov matrix M
e
symmetric matrix L.
4.3 Selecting Kernels
For demonstration purpose, we present the kernel selection details on the image of an eye
shown below. To begin with, a random walk is defined where each pixel in the test image
is associated with a vertex of the graph G. The edges in G are defined by the standard
8-neighbourhood of each pixel. For the demonstrations in this paper, the edge weight
ai,j between neighbouring pixels xi and xj is given by a function of the difference in the
corresponding intensities I(xi ) and I(xj ): ai,j = exp(?(I(xi ) ? I(xj ))2 /2?a2 ), where
?a is set according to the median absolute difference |I(xi ) ? I(xj )| between neighbours
measured over the entire image. The affinity matrix A with the edge weights is then used
to generate a Markov transition matrix M .
The kernel selection process we use is fast and greedy. First, the fine scale Markov matrix
M is diffused to M ? using ? = 4. The Markov matrix M is sparse as we make the
affinity matrix A sparse. Every column in the diffused matrix M ? is a potential kernel. To
facilitate the selection process, the second step is to rank order the columns of M ? based
on a probability value in the stationary distribution ~? . Third, the kernels (i.e. columns of
M ? ) are picked in such a way that for a kernel Ki all of the neighbours of pixel i which are
within the half-height of the the maximum value in the kernel Ki are suppressed from the
selection process. Finally, the kernel selection is continued until every pixel in the image is
within a half-height of the peak value of at least one kernel. If M is a full matrix, to avoid
the expense of computing M ? explicitly, random kernel centers can be selected, and only
the corresponding columns of M ? need be computed.
We show results from a three-scale hierarchy on the eye image (below). The image has
25 ? 20 pixels but is shown here enlarged for clarity. At the first coarse scale 83 kernels
are picked. The kernels each correspond to a different column in the fine scale transition
matrix and the pixels giving rise to these kernels are shown numbered on the image.
Using these kernels as an initialization, the EM procedure derives a coarse-scale stationary distribution ~?
21
14
26
4
(Eq. 2), while simultaneously updating the kernel ma12
27
2
19
trix. Using the newly updated kernel matrix K and the
5
8 13
23 30
f
18
6 9
derived stationary distribution ~? a transition matrix M
28
20
15
32
10
22
is
generated
(Eq.
9).
The
coarse
scale
Markov
matrix
24
17
7
f? , again using ? = 4. The kernel
is then diffused to M
Coarse Scale 1 Coarse Scale 2
selection algorithm is reapplied, this time picking 32
kernels for the second coarse scale. Larger values of ? cause the coarser level to have fewer
elements. But the exact number of elements depends on the form of the kernels themselves.
For the random experiments that we describe later in ?6 we found ? = 2 in the first iteration
and 4 thereafter causes the number of kernels to be reduced by a factor of roughly 1/3 to
1/4 at each level.
72
28 35 44 51
64 82
4 12 31
56
19
77
36 45
52 65
13
57
23 37
5
40
53 63 73
14 29
6
66
38
74
47
24
7
30 41 54 71 78
58
15
8 20
25 39 48 59 67
68 79
21
16 26 42 49
2 11
55
60 75
32
83
43
9
76
50
17 27
61
33
69 80
3
46
18
70 81
34
10
62
22
1
25
11
1
3
16
31
29
At coarser levels of the hierarchy, we expect the kernels to get less sparse and so will the
affinity and the transition matrices. In order to promote sparsity at successive levels of
e by zeroing out elements associated with ?small? transition
the hierarchy we sparsify A
f
probabilities in M . However, in the experiments described later in ?6, we observe this
sparsification step to be not critical. To summarize, we use the stationary distribution ~?
f, and its stationary distribution ~?, at the
at the fine-scale to derive a transition matrix M
e and
coarse-scale. The coarse scale transition in turn helps to derive an affinity matrix A
e It is obvious that this procedure can be repeated recursively. We
its normalized version L.
describe next how to use this representation hierarchy for building a fast eigensolver.
5 Fast EigenSolver
Our goal in generating a hierarchical representation of a transition matrix is to develop a
fast, specialized eigen solver for spectral clustering. To this end, we perform a full eigen
decomposition of the normalized affinity matrix only at the coarsest level. As discussed
in the previous section, the affinity matrix at the coarsest level is not likely to be sparse,
hence it will need a full (as opposed to a sparse) version of an eigen solver. However it is
typically the case that e ? m n (even in the case of the three-scale hierarchy that we
just considered) and hence we expect this step to be the least expensive computationally.
The resulting eigenvectors are interpolated to the next lower level of the hierarchy by a
process which will be described next. Because the eigen interpolation process between
every adjacent pair of scales in the hierarchy is similar, we will assume we have access to
e (size: m ? e) for the normalized affinity matrix L
e (size: m ? m)
the leading eigenvectors U
and describe how to generate the leading eigenvectors U (size: n ? e), and the leading
eigenvalues S (size: e ? 1), for the fine-scale normalized affinity matrix L (size: n ? n).
There are several steps to the eigen interpolation process and in the discussion that follows
we refer to the lines in the pseudo-code presented below.
e can be interpolated using the kernel matrix K to
First, the coarse-scale eigenvectors U
e
generate U = K U , an approximation for the fine-scale eigenvectors (line 9). Second,
interpolation alone is unlikely to set the directions of U exactly aligned with U L , the
vectors one would obtain by a direct eigen decomposition of the fine-scale normalized
affinity matrix L. We therefore update the directions in U by applying a small number of
power iterations with L, as given in lines 13-15.
e , S)
e
function (U, S) = CoarseToFine(L, K, U
1: INPUT
2:
L, K ? {L is n ? n and K is n ? m where m n}
e /Se ? {leading coarse-scale eigenvectors/eigenvalues of L.
e U
e is of size m ? e, e ? m}
3:
U
4: OUTPUT
5:
U, S ? {leading fine-scale eigenvectors/eigenvalues of L. U is n ? e and S is e ? 1.}
x 10
0.4
3
0.96
0.94
0.92
0.9
0.35
2.5
Relative Error
Absolute Relative Error
0.98
Eigen Value
|??|??1
?3
Eigen Spectrum
1
2
1.5
1
5
10
15
20
Eigen Index
(a)
25
30
0.2
0.15
0.1
0.5
0.88
0.3
0.25
0.05
5
10
15
20
Eigen Index
(b)
25
30
5
10
15
20
Eigen Index
25
30
(c)
Figure 1: Hierarchical eigensolver results. (a) comparing ground truth eigenvalues S L
(red circles) with multi-scale eigensolver spectrum S (blue line) (b) Relative absolute error
L|
between eigenvalues: |S?S
(c) Eigenvector mismatch: 1 ? diag |U T UL | , between
SL
eigenvectors U derived by the multi-scale eigensolver and the ground truth U L . Observe
the slight mismatch in the last few eigenvectors, but excellent agreement in the leading
eigenvectors (see text).
6: CONSTANTS: TOL = 1e-4; POWER ITERS = 50
7:
?
?
e
8: TPI = min POWER ITERS, log(e ? eps/TOL)/ log(min(S))
{eps: machine accuracy}
e
9: U = K U {interpolation from coarse to fine}
10: while not converged do
11: Uold = U {n ? e matrix, e n}
12: for i = 1 to TPI do
13:
U ? LU
14: end for
15: U ? Gram-Schmidt(U ) {orthogonalize U }
16: Le = U T LU {L may be sparse, but Le need not be.}
17: Ue Se UeT = svd(Le ) {eigenanalysis of Le , which is of size e ? e.}
18: U ? U Ue {update the leading eigenvectors of L}
19: S = diag(Se ) {grab the leading eigenvalues of L}
T
U ) {1 is a e ? 1 vector of all ones}
20: innerProd = 1 ? diag( Uold
21: converged = max[abs(innerProd)] < TOL
22: end while
The number of power iterations TPI can be bounded as discussed next. Suppose ~v =
U c where U is a matrix of true eigenvectors and c is a coefficient vector for an arbitrary
vector ~v . After TPI power iterations ~v becomes ~v = U diag(S TPI )c, where S has the exact
eigenvalues. In order for the component of a vector ~v in the direction Ue (the eth column of
U ) not to be swamped by other components, we can limit it?s decay after TPI iterations as
TPI
follows: (S(e)/S(1)) >= e?eps/TOL, where S(e) is the exact eth eigenvalue, S(1) =
1, eps is the machine precision, TOL is requested accuracy. Because we do not have access
to the exact value S(e) at the beginning of the interpolation procedure, we estimate it from
e This leads to a bound on the power iterations TPI, as derived on
the coarse eigenvalues S.
the line 9 above. Third, the interpolation process and the power iterations need not preserve
orthogonality in the eigenvectors in U . We fix this by Gram-Schmidt orthogonalization
procedure (line 16). Finally, there is a still a problem with power iterations that needs
to be resolved, in that it is very hard to separate nearby eigenvalues. In particular, for
the convergence of the power iterations the ratio that matters is between the (e + 1) st and
eth eigenvalues. So the idea we pursue is to use the power iterations only to separate the
reduced space of eigenvectors (of dimension e) from the orthogonal subspace (of dimension
n ? e). We then use a full SVD on the reduced space to update the leading eigenvectors
U , and eigenvalues S, for the fine-scale (lines 17-20). This idea is similar to computing the
Ritz values and Ritz vectors in a Rayleigh-Ritz method.
6 Interpolation Results
Our multi-scale decomposition code is in Matlab. For the direct eigen decomposition, we
have used the Matlab program svds.m which invokes the compiled ARPACKC routine
[13], with a default convergence tolerance of 1e-10.
In Fig. 1a we compare the spectrum S obtained from a three-scale decomposition on the
eye image (blue line) with the ground truth, which is the spectrum SL resulting from direct
eigen decomposition of the fine-scale normalized affinity matrices L (red circles). There
is an excellent agreement in the leading eigenvalues. To illustrate this, we show absolute
L|
in Fig. 1b. The spectra agree mostly, except for
relative error between the spectra: |S?S
SL
the last few eigenvalues. For a quantitative comparison between the eigenvectors, we plot
in Fig. 1c the following measure: 1 ? diag(|U T UL |), where U is the matrix of eigenvectors
obtained by the multi-scale approximation, UL is the ground-truth resulting from a direct
eigen decomposition of the fine-scale affinity matrix L and 1 is a vector of all ones. The
relative error plot demonstrates a close match, within the tolerance threshold of 1e-4 that
we chose for the multi-scale method, in the leading eigenvector directions between the
two methods. The relative error is high with the last few eigen vectors, which suggests
that the power iterations have not clearly separated them from other directions. So, the
strategy we suggest is to pad the required number of leading eigen basis by about 20%
before invoking the multi-scale procedure. Obviously, the number of hierarchical stages
for the multi-scale procedure must be chosen such that the transition matrix at the coarsest
scale can accommodate the slight increase in the subspace dimensions. For lack of space
we are omitting extra results (see Ch.8 in [6]).
Next we measure the time the hierarchical eigensolver takes to compute the leading eigenbasis for various input sizes, in comparison with the svds.m procedure [13]. We form
images of different input sizes by Gaussian smoothing of i.i.d noise. The Gaussian function has a standard deviation of 3 pixels. The edges in graph G are defined by the standard
8-neighbourhood of each pixel. The edge weights between neighbouring pixels are simply given by a function of the difference in the corresponding intensities (see ?4.3). The
affinity matrix A with the edge weights is then used to generate a Markov transition matrix
M . The fast eigensolver is run on ten different instances of the input image of a given size
and the average of these times is reported here. For a fair comparison between the two
procedures, we set the convergence tolerance value for the svds.m procedure to be 1e-4,
the same as the one used for the fast eigensolver. We found the hierarchical representation
derived from this tolerance threshold to be sufficiently accurate for a novel min-cut based
segmentation results that we reported in [8]. Also, the subspace dimensionality is fixed to
be 51 where we expect (and indeed observe) the leading 40 eigenpairs derived from the
multi-scale procedure to be accurate. Hence, while invoking svds.m we compute only the
leading 41 eigenpairs.
In the table shown below, the first column corresponds to the number of nodes in the graph,
while the second and third columns report the time taken in seconds by the svds.m procedure and the Matlab implementation of the multi-scale eigensolver respectively. The
fourth column reports the speedups of the multi-scale eigensolver over svds.m procedure
on a standard desktop (Intel P4, 2.5GHz, 1GB RAM). Lowering the tolerance threshold for
svds.m made it faster by about 20 ? 30%. Despite this, the multi-scale algorithm clearly
outperforms the svds.m procedure. The most expensive step in the multi-scale algorithm
is the power iteration required in the last stage, that is interpolating eigenvectors from the
first coarse scale to the required fine scale. The complexity is of the order of n ? e where
e is the subspace dimensionality and n is the size of the graph. Indeed, from the table we
can see that the multi-scale procedure is taking time roughly proportional to n. Deviations
from the linear trend are observed at specific values of n, which we believe are due to the
n
322
632
642
652
1002
1272
1282
1292
1602
2552
2562
2572
5112
5122
5132
6002
7002
8002
svds.m
1.6
10.8
20.5
12.6
44.2
91.1
230.9
96.9
179.3
819.2
2170.8
871.7
7977.2
20269
7887.2
10841.4
15048.8
Multi-Scale
1.5
4.9
5.5
5.1
13.1
20.4
35.2
20.9
34.4
90.3
188.7
93.3
458.8
739.3
461.9
644.2
1162.4
1936.6
Speedup
1.1
2.2
3.7
2.5
3.4
4.5
6.6
4.6
5.2
9.1
11.5
9.3
17.4
27.4
17.1
16.8
12.9
variations in the difficulty of the specific eigenvalue problem (eg. nearly multiple eigenvalues).
The hierarchical representation has proven to be
effective in a min-cut based segmentation algorithm that we proposed recently [8]. Here we explored the use of random walks and associated
spectral embedding techniques for the automatic
generation of suitable proposal (source and sink)
regions for a min-cut based algorithm. The multiscale algorithm was used to generate the 40 leading eigenvectors of large transition matrices (eg.
size 20K ? 20K). In terms of future work, it will
be useful to compare our work with other approximate methods for SVD such as [23].
Ack: We thank S. Roweis, F. Estrada and M. Sakr for valuable comments.
References
[1] D. Achlioptas and F. McSherry. Fast Computation of Low-Rank Approximations. STOC, 2001.
[2] D. Achlioptas et al Sampling Techniques for Kernel Methods. NIPS, 2001.
[3] S. Barnard and H. Simon Fast Multilevel Implementation of Recursive Spectral Bisection for
Partitioning Unstructured Problems. PPSC, 627-632.
[4] M. Belkin et al Laplacian Eigenmaps and Spectral Techniques for Embedding. NIPS, 2001.
[5] M. Brand et al A unifying theorem for spectral embedding and clustering. AI & STATS, 2002.
[6] C. Chennubhotla. Spectral Methods for Multi-scale Feature Extraction and Spectral Clustering.
http://www.cs.toronto.edu/?chakra/thesis.pdf Ph.D Thesis, Department of
Computer Science, University of Toronto, Canada, 2004.
[7] C. Chennubhotla and A. Jepson. Half-Lives of EigenFlows for Spectral Clustering. NIPS, 2002.
[8] F. Estrada, A. Jepson and C. Chennubhotla. Spectral Embedding and Min-Cut for Image Segmentation. Manuscript Under Review, 2004.
[9] C. Fowlkes et al Efficient spatiotemporal grouping using the Nystrom method. CVPR, 2001.
[10] S. Belongie et al Spectral Partitioning with Indefinite Kernels using Nystrom app. ECCV,
2002.
[11] A. Frieze et al Fast Monte-Carlo Algorithms for finding low-rank approximations. FOCS,
1998.
[12] Y. Koren et al ACE: A Fast Multiscale Eigenvectors Computation for Drawing Huge Graphs
IEEE Symp. on InfoVis 2002, pp. 137-144
[13] R. B. Lehoucq, D. C. Sorensen and C. Yang. ARPACK User Guide: Solution of Large Scale
Eigenvalue Problems by Implicitly Restarted Arnoldi Methods. SIAM 1998.
[14] J. J. Lin. Reduced Rank Approximations of Transition Matrices. AI & STATS, 2002.
[15] L. Lova?sz. Random Walks on Graphs: A Survey Combinatorics, 1996, 353?398.
[16] G. J. McLachlan et al Mixture Models: Inference and Applications to Clustering. 1988
[17] M. Meila and J. Shi. A random walks view of spectral segmentation. AI & STATS, 2001.
[18] A. Ng, M. Jordan and Y. Weiss. On Spectral Clustering: analysis and an algorithm NIPS, 2001.
[19] A. Pothen Graph partitioning algorithms with applications to scientific computing. Parallel
Numerical Algorithms, D. E. Keyes et al (eds.), Kluwer Academic Press, 1996.
[20] G. L. Scott et al Feature grouping by relocalization of eigenvectors of the proximity matrix.
BMVC, pg. 103-108, 1990.
[21] E. Sharon et al Fast Multiscale Image Segmentation CVPR, I:70-77, 2000.
[22] J. Shi and J. Malik. Normalized cuts and image segmentation. PAMI, August, 2000.
[23] H. Simon et al Low-Rank Matrix Approximation Using the Lanczos Bidiagonalization Process
with Applications SIAM J. of Sci. Comp. 21(6):2257-2274, 2000.
[24] N. Tishby et al Data clustering by Markovian Relaxation NIPS, 2001.
[25] C. Williams et al Using the Nystrom method to speed up the kernel machines. NIPS, 2001.
| 2665 |@word version:2 decomposition:12 pg:1 invoking:2 pick:2 accommodate:1 recursively:1 ld:1 initial:1 selecting:1 pna:1 diagonalized:1 outperforms:1 comparing:2 must:1 numerical:1 plot:2 update:4 stationary:15 pursued:1 greedy:2 half:3 de1:1 selected:1 fewer:1 desktop:1 alone:1 beginning:1 ith:2 qjk:2 coarse:27 node:8 toronto:3 successive:5 height:2 dn:1 direct:5 focs:1 symp:1 pairwise:1 allan:1 indeed:2 roughly:2 themselves:1 multi:15 decreasing:1 considering:2 solver:3 becomes:1 begin:2 notation:1 bounded:1 pursue:1 eigenvector:6 finding:1 transformation:1 sparsification:1 pseudo:1 quantitative:1 every:4 exactly:1 demonstrates:1 uk:1 partitioning:3 unit:1 arnoldi:1 segmenting:1 before:1 eigenpairs:2 understood:1 local:1 limit:1 despite:1 analyzing:1 interpolation:7 pami:1 chose:1 initialization:3 suggests:1 ease:1 practical:1 unique:1 recursive:1 procedure:18 eth:3 numbered:1 suggest:1 get:3 close:1 selection:7 applying:1 instability:1 www:1 lagrangian:1 center:1 shi:2 williams:1 starting:1 survey:1 unstructured:1 stats:3 continued:1 ritz:3 deriving:2 de2:1 embedding:5 variation:1 updated:1 hierarchy:14 suppose:1 user:1 exact:5 neighbouring:2 agreement:2 lova:1 element:4 trend:1 expensive:4 updating:1 cut:6 coarser:2 observed:1 solved:2 svds:9 region:1 connected:1 valuable:1 complexity:1 basis:1 sink:1 easily:1 resolved:1 represented:3 various:1 separated:1 fast:17 effective:2 describe:3 monte:1 lift:1 choosing:1 whose:1 ace:2 larger:1 solve:2 cvpr:2 say:1 drawing:1 think:1 obviously:1 eigenvalue:26 tpi:8 propose:1 p4:1 neighboring:1 aligned:1 representational:1 roweis:1 eigenbasis:1 convergence:3 cluster:1 generating:1 help:2 derive:4 develop:2 illustrate:1 measured:1 eq:6 c:1 eigenflows:1 differ:1 direction:5 undertaking:1 closely:1 correct:2 stochastic:5 enable:1 multilevel:1 fix:1 proximity:1 sufficiently:1 considered:1 ground:4 exp:1 substituting:1 a2:1 purpose:1 largest:1 mclachlan:1 clearly:2 gaussian:3 pn:1 avoid:1 sparsify:2 derived:6 notational:1 rank:7 likelihood:1 dim:1 inference:1 typically:2 entire:1 unlikely:1 pad:1 expand:1 pixel:10 classification:1 smoothing:1 iters:2 equal:1 construct:1 never:2 eigenproblem:1 having:1 sampling:2 extraction:1 biology:1 represents:1 ng:1 nearly:1 promote:1 future:1 report:2 few:3 belkin:1 randomly:1 neighbour:2 frieze:1 simultaneously:2 preserve:2 ab:1 huge:1 mixture:3 mcsherry:1 sorensen:1 chain:5 uold:2 accurate:2 integral:1 edge:7 necessary:1 orthogonal:2 walk:10 circle:2 instance:1 column:16 markovian:4 lanczos:1 vertex:3 deviation:2 uniform:1 eigenmaps:1 pothen:1 tishby:1 reported:3 spatiotemporal:1 st:1 peak:1 siam:2 probabilistic:1 picking:1 again:1 thesis:2 successively:2 opposed:1 leading:18 potential:1 dem:1 coefficient:2 matter:1 combinatorics:1 explicitly:4 vi:2 ad:2 depends:1 later:3 break:1 picked:2 view:1 red:2 parallel:2 defer:1 simon:2 accuracy:3 bidiagonalization:1 correspond:1 bisection:1 lu:2 carlo:1 comp:1 app:1 converged:2 ed:2 definition:2 pp:1 obvious:1 nystrom:5 naturally:1 associated:3 di:1 newly:1 dataset:1 reapplied:1 dimensionality:2 segmentation:7 routine:1 back:1 chakra:2 manuscript:1 originally:1 wei:1 bmvc:1 formulation:1 generality:1 just:2 stage:4 implicit:1 achlioptas:2 until:1 ei:1 multiscale:3 propagation:1 lack:1 defines:1 scientific:2 believe:1 facilitate:1 omitting:1 building:3 normalized:9 true:1 analytically:3 hence:3 symmetric:6 eg:2 adjacent:1 ue:3 acquires:1 ack:1 pdf:1 theoretic:1 complete:1 orthogonalization:1 image:15 novel:1 recently:3 specialized:3 fdiag:1 overview:1 extend:1 discussed:3 slight:2 kluwer:1 eps:4 measurement:1 expressing:1 refer:1 ai:8 automatic:1 meila:1 grid:1 pm:2 zeroing:1 particle:2 dj:1 access:2 similarity:2 compiled:1 chennubhotla:4 life:1 preceding:1 estrada:2 employed:1 full:5 multiple:1 match:1 faster:1 academic:1 lin:2 laplacian:1 regression:1 iteration:16 kernel:34 proposal:1 eigensolver:15 fine:18 median:1 source:1 extra:1 comment:1 undirected:1 coarsening:2 jordan:1 yang:1 easy:3 variety:1 xj:4 idea:4 expression:1 gb:1 ul:3 cause:3 matlab:4 tol:5 useful:1 se:3 eigenvectors:26 relocalization:1 ten:2 ph:1 reduced:6 generate:11 http:1 sl:3 popularity:1 blue:2 write:1 thereafter:1 indefinite:1 terminology:2 dej:1 threshold:3 achieving:1 clarity:1 diffusion:1 lowering:1 sharon:1 ram:1 graph:18 relaxation:4 grab:1 sum:2 run:1 fourth:1 pik:1 ki:9 bound:1 koren:1 precisely:1 orthogonality:1 ri:3 nearby:1 interpolated:4 u1:1 speed:1 min:7 coarsest:5 circumvented:1 speedup:3 department:3 according:1 em:5 suppressed:1 partitioned:1 swamped:1 taken:2 computationally:2 ln:1 equation:1 agree:1 turn:2 end:3 keyes:1 apply:1 observe:4 hierarchical:11 spectral:19 fowlkes:1 neighbourhood:2 schmidt:2 eigen:27 substitute:1 clustering:11 remaining:1 include:1 unifying:1 exploit:1 giving:1 invokes:1 build:1 objective:1 diffused:4 malik:1 strategy:1 diagonal:1 lehoucq:1 affinity:24 subspace:5 link:1 separate:2 thank:1 sci:1 reason:1 assuming:1 length:2 code:2 index:3 ratio:1 demonstration:2 mostly:1 potentially:1 stoc:1 expense:1 negative:4 rise:1 design:1 implementation:3 unknown:1 perform:3 markov:17 datasets:1 qtk:2 perturbation:1 arbitrary:1 august:1 intensity:2 canada:1 namely:1 pair:1 required:3 nip:6 below:5 mismatch:2 scott:1 sparsity:1 summarize:1 program:1 max:1 power:14 suitable:2 critical:2 difficulty:2 ranked:2 brief:1 eye:3 carried:2 speeding:1 text:1 review:1 relative:6 loss:1 expect:3 generation:1 proportional:1 proven:2 degree:3 sufficient:1 eccv:1 last:4 guide:1 taking:1 emerge:1 absolute:4 sparse:7 tolerance:5 ghz:1 dimension:3 default:1 transition:29 kdiag:1 gram:2 made:1 approximate:4 preferred:1 arpack:1 implicitly:1 dealing:1 sz:1 global:2 pittsburgh:1 belongie:1 xi:4 spectrum:6 un:1 latent:3 table:2 requested:1 excellent:2 interpolating:1 domain:1 vj:1 jepson:3 diag:14 main:1 noise:1 repeated:1 fair:1 enlarged:1 fig:3 representative:1 intel:1 precision:1 wish:1 third:3 down:1 theorem:1 specific:2 explored:1 decay:1 eigenanalysis:1 derives:1 grouping:2 gained:1 rayleigh:1 distinguishable:1 simply:1 likely:1 expressed:2 trix:1 u2:1 restarted:1 ch:1 corresponds:2 truth:4 goal:4 sorted:1 ownership:2 barnard:1 hard:1 typical:2 specifically:1 reducing:1 except:1 svd:4 orthogonalize:1 brand:1 select:1 d1:5 correlated:1 |
1,834 | 2,666 | An Investigation of Practical Approximate
Nearest Neighbor Algorithms
Ting Liu, Andrew W. Moore, Alexander Gray and Ke Yang
School of Computer Science
Carnegie-Mellon University
Pittsburgh, PA 15213 USA
{tingliu, awm, agray, yangke}@cs.cmu.edu
Abstract
This paper concerns approximate nearest neighbor searching algorithms,
which have become increasingly important, especially in high dimensional perception areas such as computer vision, with dozens of publications in recent years. Much of this enthusiasm is due to a successful new
approximate nearest neighbor approach called Locality Sensitive Hashing (LSH). In this paper we ask the question: can earlier spatial data
structure approaches to exact nearest neighbor, such as metric trees, be
altered to provide approximate answers to proximity queries and if so,
how? We introduce a new kind of metric tree that allows overlap: certain
datapoints may appear in both the children of a parent. We also introduce new approximate k-NN search algorithms on this structure. We
show why these structures should be able to exploit the same randomprojection-based approximations that LSH enjoys, but with a simpler algorithm and perhaps with greater efficiency. We then provide a detailed
empirical evaluation on five large, high dimensional datasets which show
up to 31-fold accelerations over LSH. This result holds true throughout
the spectrum of approximation levels.
1
Introduction
The k-nearest-neighbor searching problem is to find the k nearest points in a dataset X ?
RD containing n points to a query point q ? RD , usually under the Euclidean distance. It
has applications in a wide range of real-world settings, in particular pattern recognition,
machine learning [7] and database querying [11]. Several effective methods exist for this
problem when the dimension D is small (e.g. 1 or 2), such as Voronoi diagrams [26],
or when the dimension is moderate (i.g. up to the 10?s), such as kd-trees [8] and metric
trees. Metric trees [29], or ball-trees [24], so far represent the practical state of the art for
achieving efficiency in the largest dimensionalities possible [22, 6]. However, many realworld problems are posed with very large dimensionalities which are beyond the capability
of such search structures to achieve sub-linear efficiency, for example in computer vision,
in which each pixel of an image represents a dimension. Thus, the high-dimensional case
is the long-standing frontier of the nearest-neighbor problem.
Approximate searching. One approach to dealing with this apparent intractability has
been to define a different problem, the (1 + ?) approximate k-nearest-neighbor searching
problem, which returns points whose distance from the query is no more than (1 + ?) times
the distance of the true kth nearest-neighbor. Further, the problem is often relaxed to only
do this with high probability, and without a certificate property telling the user when it
has failed to do so, nor any guarantee on the actual rank of the distance of the points
returned, which may be arbitrarily far from k [4]. Another commonly used modification to
the problem is to perform the search under the L1 norm rather than L2 .
Locality Sensitive Hashing. Several methods of this general nature have been proposed
[17, 18, 12], and locality-sensitive hashing (LSH) [12] has received considerable recent
attention because it was shown that its runtime is independent of the dimension D and
has been put forth as a practical tool [9]. Roughly speaking, a locality sensitive hashing
function has the property that if two points are ?close,? then they hash to same bucket
with ?high? probability; if they are ?far apart,? then they hash to same bucket with ?low?
probability. Formally, a function family H = {h : S ? U} is (r1 , r2 , p1 , p2 )-sensitive, where
r1 < r2 , p1 > p2, for distance function D if for any two points p, q ? S, the following
properties hold:
1. if p ? B(q, r1 ), then Prh?H [h(q) = h(p)] ? p1 , and
2. if p 6? B(q, r2 ), then Prh?H [h(q) = h(p)] ? p2 ,
where B(q, r) denotes a hypersphere of radius r centered at q. By defining a LSH scheme,
namely a (r, r(1 + ?), p1 , p2 )-sensitive hash family, the (1 + ?)-NN problem can be solved
by performing a series of hashing and searching within the buckets. See [12, 13] for details.
Applications such as computer vision, e.g. [23, 28] have found (1 + ?) approximation to
be useful, for example when the k-nearest-neighbor search is just one component in a large
system with many parts, each of which can be highly inaccurate. In this paper we explore
the extent to which the most successful exact search structures can be adapted to perform
(1 + ?) approximate high-dimensional searches. A notable previous approach along this
line is a simple modification of kd-trees [3] ? ours takes the more powerful metric trees as
a starting point. We next review metric trees, then introduce a variant, known as spill trees.
2
Metric Trees and Spill Trees
2.1 Metric Trees
The metric tree [29, 25, 5] is a data structure that supports efficient nearest neighbor search.
We briefly A metric tree organizes a set of points in a spatial hierarchical manner. It is a
binary tree whose nodes represent a set of points. The root node represents all points, and
the points represented by an internal node v is partitioned into two subsets, represented by
its two children. Formally, if we use N(v) to denote the set of points represented by node
v, and use v.lc and v.rc to denote the left child and the right child of node v, then we have
N(v) = N(v.lc) ? N(v.rc)
(1)
0/ = N(v.lc) ? N(v.rc)
(2)
for all the non-leaf nodes. At the lowest level, each leaf node contains very few points.
Partitioning. The key to building a metric-tree is how to partition a node v. A typical
way is as follows. We first choose two pivot points from N(v), denoted as v.lpv and v.rpv.
Ideally, v.lpv and v.rpv are chosen so that the distance between them is the largest of allpair distances within N(v). More specifically, ||v.lpv ? v.rpv|| = max p1,p2?N(v) ||p1 ? p2||.
However, it takes O(n2 ) time to find the optimal v.lpv and v.rpv. In practice, we resort to a
linear-time heuristic that is still able to find reasonable pivot points.1 After v.lpv and v.rpv
are found, we can go ahead to partition node v.
Here is one possible strategy for partitioning. We first project all the points down to the
~
~ ? v.lpv,
vector ~u = v.rpv
and then find the median point A along ~u. Next, we assign all
the points projected to the left of A to v.lc, and all the points projected to the right of A
to v.rc. We use L to denote the d-1 dimensional plane orthogonal to ~u and goes through
A. It is known as the decision boundary since all points to the left of L belong to v.lc
1 Basically, we first randomly pick a point p from v. Then we search for the point that is the
farthest to p and set it to be v.lpv. Next we find a third point that is farthest to v.lpv and set it as v.rpv.
Figure 1: partitioning in a metric tree.
Figure 2: partitioning in a spill tree.
and all points to the right of L belong to v.rc (see Figure 1). By using a median point to
split the datapoints, we can ensure that the depth of a metric-tree is log n. However, in our
~ + v.rpv))
~
implementation, we use a mid point (i.e. the point at 12 (v.lpv
instead, since it is
more efficient to compute, and in practice, we can still have a metric tree of depth O(log n).
Each node v also has a hypersphere B, such that all points represented by v fall in the ball
centered at v.center with radius v.r, i.e. we have N(v) ? B(v.center, v.r). Notice that the
balls of the two children nodes are not necessarily disjoint.
Searching. A search on a metric-tree is simply a guided DFS (for simplicity, we assume
that k = 1). The decision boundary L is used to decide which child node to search first. If
the query q is on the left of L, then v.lc is searched first, otherwise, v.rc is searched first. At
all times, the algorithm maintains a ?candidate NN?, which is the nearest neighbor it finds
so far while traversing the tree. We call this point x, and denote the distance between q and
x by r. If DFS is about to exploit a node v, but discovers that no member of v can be within
distance r of q, then it prunes this node (i.e., skip searching on this node, along with all its
descendants). This happens whenever kv.center ? qk ? v.r ? r. We call this DFS search
algorithm MT-DFS thereafter.
In practice, the MT-DFS algorithm is very efficient for NN search, and particularly when
the dimension of a dataset is low (say, less than 30). Typically for MT-DFS, we observe
an order of magnitude speed-up over na??ve linear scan and other popular data structures
such as SR-trees. However, MT-DFS starts to slow down as the dimension of the datasets
increases. We have found that in practice, metric tree search typically finds a very good NN
candidate quickly, and then spends up to 95% of the time verifying that it is in fact the true
NN. This motivated our new proposed structure, the spill-tree, which is designed to avoid
the cost of exact NN verification.
2.2 Spill-Trees
A spill-tree (sp-tree) is a variant of metric-trees in which the children of a node can ?spill
over? onto each other, and contain shared datapoints. The partition procedure of a metrictree implies that point-sets of v.lc and v.rc are disjoint: these two sets are separated by the
decision boundary L. In a sp-tree, we change the splitting criteria to allow overlaps between
two children. In other words, some datapoints may belong to both v.lc and v.rc.
We first explain how to split an internal node v. See Figure 2 as an example. Like a metrictree, we first choose two pivots v.lpv and v.rpv, and find the decision boundary L that goes
through the mid point A, Next, we define two new separating planes, LL and LR, both of
which are parallel to L and at distance ? from L. Then, all the points to the right of plane
LL belong to the child v.rc, and all the points to the left of plane LR belong to the child v.lc.
Mathematically, we have
N(v.lc) = {x | x ? N(v), d(x, LR) + 2? > d(x, LL)}
N(v.rc) = {x | x ? N(v), d(x, LL) + 2? > d(x, LR)}
(3)
(4)
Notice that points fall in the region between LL and LR are shared by v.lc and v.rc. We call
this region the overlapping buffer, and we call ? the overlapping size. For v.lc and v.rc, we
can repeat the splitting procedure, until the number of points within a node is less than a
specific threshold, at which point we stop.
3
Approximate Spill-tree-based Nearest Neighbor Search
It may seem strange that we allow overlapping in sp-trees. The overlapping obviously
makes both the construction and the MT-DFS less efficient than regular metric-trees, since
the points in the overlapping buffer may be searched twice. Nonetheless, the advantage of
sp-trees over metric-trees becomes clear when we perform the defeatist search, an (1 + ?)NN search algorithm based on sp-trees.
3.1 Defeatist Search
As we have stated, the MT-DFS algorithm typically spends a large fraction of time backtracking to prove a candidate point is the true NN. Based on this observation, a quick
revision would be to descends the metric tree using the decision boundaries at each level
without backtracking, and then output the point x in the first leaf node it visits as the NN of
query q. We call this the defeatist search on a metric-tree. Since the depth of a metric-tree
is O(log n), the complexity of defeatist search is O(log n) per query.
The problem with this approach is very low accuracy. Consider the case where q is very
close to a decision boundary L, then it is almost equally likely that the NN of q is on the
same side of L as on the opposite side of L, and the defeatist search can make a mistake
with probability close to 1/2. In practice, we observe that there exists a non-negligible
fraction of the query points that are close to one of the decision boundaries. Thus the
average accuracy of the defeatist search algorithm is typically unacceptably low, even for
approximate NN search.
This is precisely the place where sp-trees can help: the defeatist search on sp-trees has much
higher accuracy and remains very fast. We first describe the algorithm. For simplicity, we
continue to use the example shown in Figure 2. As before, the decision boundary at node
v is plane L. If a query q is to the left of L, we decide that its nearest neighbor is in v.lc. In
this case, we only search points within N(v.lc), i.e., the points to the left of LR. Conversely,
if q is to the right of L, we only search node v.rc, i.e. points to the right of LL. Notice that
in either case, points in the overlapping buffer are always searched. By introducing this
buffer of size ?, we can greatly reduce the probability of making a wrong decision. To see
this, suppose that q is to the left of L, then the only points eliminated are the one to the right
of plane LR, all of which are at least distance ? away from q.
3.2 Hybrid Sp-Tree Search
One problem with spill-trees is that their depth varies considerably depending on the overlapping size ?. If ? = 0, a sp-tree turns back to a metric tree with depth O(log n). On the
other hand, if ? ? ||v.rpv ? v.lpv||/2, then N(v.lc) = N(v.rc) = N(v). In other words, both
children of node v contain all points of v. In this case, the construction of a sp-tree does
not even terminate and the depth of the sp-tree is ?.
To solve this problem, we introduce hybrid sp-trees and actually use them in practice. First
we define a balance threshold ? < 1, which is usually set to 70%. The constructions of
a hybrid sp-tree is similar to that of a sp-tree except the following. For each node v, we
first split the points using the overlapping buffer. However, if either of its children contains
more than ? fraction of the total points in v, we undo the overlapping splitting. Instead, a
conventional metric-tree partition (without overlapping) is used, and we mark v as a nonoverlapping node. In contrast, all other nodes are marked as overlapping nodes. In this
way, we can ensure that each split reduces the number of points of a node by at least a
constant factor and thus we can maintain the logarithmic depth of the tree.
The NN search on a hybrid sp-tree also becomes a hybrid of the MT-DFS search and the
defeatist search. We only do defeatist search on overlapping nodes, for non-overlapping
nodes, we still do backtracking as MT-DFS search. Notice that we can control the hybrid
by varying ?. If ? = 0, we have a pure sp-tree with defeatist search ? very efficient but
not accurate enough; if ? ? ||v.rpv ? v.lpv||/2, then every node is a non-overlapping node
(due to the balance threshold mechanism) ? in this way we get back to the traditional
metric-tree with MT-DFS, which is perfectly accurate but inefficient. By setting ? to be
somewhere in between, we can achieve a balance of efficiency and accuracy. As a general
rule, the greater ? is, the more accurate and the slower the search algorithm becomes.
3.3 Further Efficiency Improvement Using Random Projection
The hybrid sp-tree search algorithm is much more efficient than the traditional MT-DFS
algorithm. However, this speed-up becomes less pronounced when the dimension of a
dataset becomes high (say, over 30). In some sense, the hybrid sp-tree search algorithm
also suffer from the curse of dimensionality, only much less severely than MT-DFS.
However, a well-known technique, namely, random projection is readily available to deal
with the high-dimensional datasets. In particular, the Johnson-Lindenstrauss Lemma [15]
states that one can embed a dataset of n points in a subspace of dimension O(log n) with little distortion on the pair-wise distances. Furthermore, the embedding is extremely simple:
one simply picks a random subspace S and project all points to S.
In our (1 + ?)-NN search algorithm, we use random projection as a pre-processing step:
project the datapoints to a subspace of lower dimension, and then do the hybrid sptree search. Both the construction of sp-tree and the search are conducted in the lowdimensional subspace. Naturally, by doing random projection, we will lose some accuracy.
But we can easily fix this problem by doing multiple rounds of random projections and doing one hybrid sp-tree search for each round. Assume the failure probability of each round
is ?, then by doing L rounds, we drive down this probability to ?L .
The core idea of the hash function used in [9] can be viewed as a variant of random projection.2 Random projection can be used as a pre-processing step in conjunction with other
? conducted a series of experiments which
techniques such as conventional MT-DFSWe
show that a modest speed-up is obtained by using random projection with MT-DFS (about
4-fold), but greater (up to 700-fold) speed-up when used with sp-tree search. Due to limited
space these results will appear in the full version of this paper [19].
4
Experimental Results
We report our experimental results based on hybrid sp-tree search on a variety of real-world
datasets, with the number of datapoints ranging from 20,000 to 275,465, and dimensions
from 60 to 3,838. The first two datasets are same as the ones used in [9], where it is
demonstrated that LSH can have a significant speedup over SR-trees.
Aerial Texture feature data contain 275,465 feature vectors of 60 dimensions representing texture
information of large aerial photographs [21, 20].
Corel hist 20,000 histograms (64-dimensional) of color thumbnail-sized images taken from the
COREL STOCK PHOTO library. However, of the 64 dimensions, only 44 of them contain non-zero entries. See [27] for more discussions. We are unable to obtain the original
dataset used in [9] from the authors, and so we reproduced our own version, following their
description. We expect that the two datasets to be almost identical.
Corel uci 68,040 histograms (64-dimensional) of color images from the COREL library. This
dataset differs significantly from Corel hist and is available from the UCI repository [1].
Disk trace 40,000 content traces of disk-write operations, each being a 1 Kilo-Byte block (therefore
having dimension 1,024). The traces are generated from a desktop computer running SuSe
Linux during daily operation.
Galaxy Spectra of 40,000 galaxies from the Sloan Digital Sky Survey, with 4000 dimensions.
Besides the sp-tree search algorithm, we also run a number of other algorithms:
LSH The original LSH implementation used in [9] is not public and we were unable to
obtain it from the authors. So we used our own efficient implementation. Experiments (described later) show that ours is comparable to the one in [9].
Na??ve The na??ve linear-scan algorithm.
2 The Johnson-Lindenstrauss Lemma only works for L norm. The ?random sampling? done in
2
the LSH in [9] roughly corresponds to the L1 version of the Johnson-Lindenstrauss Lemma.
SR-tree We use the implementation of SR-trees by Katayama and Satoh [16].
Metric-Tree This is highly optimized k-NN search based on metric trees [29, 22], and
code is publicly available [2].
The experiments are run on a dual-processor AMD Opteron machine of 1.60 GHz with
8 GB RAM. We perform 10-fold cross-validation on all the datasets. We measure the
CPU time and accuracy of each algorithm. Since all the experiments are memory-based
(all the data fit into memory completely), there is no disk access during our experiments.
To measure accuracy, we use the effective distance error [3, 9], which is defined as
d
E = Q1 ?q?Q dalg
? ? 1 , where dalg is the distance from a query q to the NN found by the
algorithm, and d ? is the distance from q to the true NN. The sum is taken over all queries.
For the k-NN case where (k > 1), we measure separately the distance ratios between the
closest points found to the nearest neighbor, the 2nd closest one to the 2nd nearest neighbor
and so on, and then take the average. Obviously, for all exact k-NN algorithms, E = 0, for
all approximate algorithms, E ? 0.
4.1 The Experiments
First, as a benchmark, we run the Na??ve, SR-tree, and the Metric-tree algorithms. All of
them find exact NN. The results are summarized in Table 1.
Table 1: the CPU time of exact SR-tree, Metric-tree, and Na??ve search
Algorithm
(%)
Naive
SR-tree
Metric-tree
Aerial
43620
23450
3650
Corel hist
(k = 1) (k = 10)
462
465
184
330
58.4
91.2
Corel uci
Disk trace
Galaxy
5460
3230
791
27050
n/a
19860
46760
n/a
6600
All the datasets are rather large, and the metric-tree is consistently the fastest. On the other
hand, the SR-tree implement only has limited speedup over the Na??ve algorithm, and it fails
to run on Disk trace and Galaxy, both of which have very high dimensions.
Then, for approximate NN search, we compare sp-tree with three other algorithms: LSH,
traditional Metric-tree and SR-tree. For each algorithm, we measure the CPU time needed
for the error E to be 1%, 2%, 5%, 10% and 20%, respectively. Since Metric-tree and SRtree are both designed for exact NN search, we also run them on randomly chosen subsets
of the whole dataset to produce approximate answers. We show the comparison results of
all algorithms for the Aerial and the Corel hist datasets, both for k = 1, in Figure 3. We
also examine the speed-up of sp-tree over other algorithms. In particular, the CPU time and
the speedup of sp-tree search over LSH is summarized in Table 2.
Aerial (D=60, n=275,476, k=1)
25000
15000
CPU time(s)
Sp-tree
LSH
Metric-tree
SR-tree
20000
CPU time(s)
Corel_hist (D=64, n=20,000, k=1)
10000
5000
0
12
5
10
Error (%)
20
400
350
300
250
200
150
100
50
0
Sp-tree
LSH
Metric-tree
SR-tree
1 2
5
10
Error (%)
20
Figure 3: CPU time (s) vs. Error (%) for selected datasets.
Since we used our own implementation of LSH, we first need to verify that it has comparable performance as the one used in [9]. We do so by examining the speedup of both
implementations over SR-tree on the Aerial and Corel hist datasets, with both k = 1 and
Table 2: the CPU time (s) of Sp-tree and its speedup (in parentheses) over LSH
Error
(%)
20
10
5
2
1
Aerial
33.5 (31)
73.2 (27)
138 (31)
286 (26)
426 (23)
Corel
(k = 1)
1.67 (8.3)
2.79 (9.8)
4.5 (11)
8 (9.5)
13.5 (6.4)
hist
(k = 10)
3.27 (6.3)
5.83 (7.0)
9.58 (6.8)
20.6 (4.2)
27.9 (4.1)
Corel uci
Disk trace
Galaxy
8.7 (8.0)
19.1 (4.9)
33.1 (4.8)
61.9 (4.4)
105 (4.1)
13.2 (5.3)
43.1 (2.9)
123 (3.7)
502 (2.5)
1590 (3.2)
24.9 (5.5)
44.2 (7.8)
76.9 (11)
110 (14)
170 (12)
k = 10 for the latter.3 For the Aerial dataset, in the case where E varies from 10% to 20%,
the speedup of LSH in [9] over SR-tree varies from 4 to 6, and as for our implementation,
the speedup varies from 4.5 to 5.4. For Corel hist, when E ranges from 2% to 20%, in
the case k = 1, the speedups of LSH in [9] ranges from 2 to 7, ours from 2 to 13. In the
case k = 10, the speedup in [9] is from 3 to 12, and ours from 4 to 16. So overall, our
implementation is comparable to, and often outperforms the one in [9].
Perhaps a little surprisingly, the Metric-tree search algorithm (MT-DFS) performs very
well on Aerial and Corel hist datasets. In both cases, when the E is small (1%), MTDFS outperforms LSH by a factor of up to 2.7, even though it aims at finding the exact
NN, while LSH only finds an approximate NN. Furthermore, the approximate MT-DFS
algorithm (conventional metric-tree based search using a random subset of the training
data) consistently outperforms LSH across the entire error spectrum on Aerial. We believe
that it is because that in both datasets, the intrinsic dimensions are quite low and thus the
Metric-tree does not suffer from the curse of dimensionality.
For the rest of the datasets, namely, Corel uci, Disk trace, and Galaxy, metric-tree becomes
rather inefficient because of the curse of dimensionality, and LSH becomes competitive.
But in all cases, sp-tree search remains the fastest among all algorithms, frequently achieving 2 or 3 orders of magnitude in speed-up. Space does not permit a lengthy conclusion, but
the summary of this paper is that there is empirical evidence that with appropriate redesign
of the data structures and search algorithms, spatial data structures remain a useful tool in
the realm of approximate k-NN search.
5
Related Work
The idea of defeastist search, i.e., non-backtracking search, has been explored by various
researchers in different contexts. See, for example, Goldstein and Ramakrishnan [10],
Yianilos [30], and Indyk [14]. The latter also proposed a data structure similar to the spilltree, where the decision boundary needs to be aligned with a coordinate and there is no
hybrid version. Indyk proved how this data structure can be used to solve approximate NN
in the L? norm.
References
[1] http://kdd.ics.uci.edu/databases/CorelFeatures/CorelFeatures.data.html.
[2] http://www.autonlab.org/autonweb/showsoftware/154/.
[3] S. Arya, D. Mount, N. Netanyahu, R. Silverman, and A. Wu. An optimal algorithm for approximate nearest neighbor searching fixed dimensions. Journal of the ACM, 45(6):891?923,
1998.
[4] Kevin Beyer, Jonathan Goldstein, Raghu Ramakrishnan, and Uri Shaft. When is ?nearest neighbor? meaningful? Lecture Notes in Computer Science, 1540:217?235, 1999.
[5] P. Ciaccia, M. Patella, and P. Zezula. M-tree: An efficient access method for similarity search
in metric spaces. In Proceedings of the 23rd VLDB International Conference, September 1997.
3 The comparison in [9] is on disk access while we compare CPU time. So strictly speaking, these
results are not comparable. Nonetheless we expect them to be more or less consistent.
[6] K. Clarkson. Nearest Neighbor Searching in Metric Spaces: Experimental Results for sb(S). ,
2002.
[7] R. O. Duda and P. E. Hart. Pattern Classification and Scene Analysis. John Wiley & Sons,
1973.
[8] J. H. Friedman, J. L. Bentley, and R. A. Finkel. An algorithm for finding best matches in logarithmic expected time. ACM Transactions on Mathematical Software, 3(3):209?226, September
1977.
[9] A. Gionis, P. Indyk, and R. Motwani. Similarity Search in High Dimensions via Hashing. In
Proc 25th VLDB Conference, 1999.
[10] J. Goldstein and R. Ramakrishnan. Constrast Polots and P-Sphere Trees: Speace vs. Time in
Nearest Neighbor Searches. In Proc. 26th VLDB conference, 2000.
[11] A. Guttman. R-trees: A dynamic index structure for spatial searching. In Proceedings of
the Third ACM SIGACT-SIGMOD Symposium on Principles of Database Systems. Assn for
Computing Machinery, April 1984.
[12] P. Indyk and R. Motwani. Approximate nearest neighbors: towards removing the curse of
dimensionality. In STOC, pages 604?613, 1998.
[13] Piotr Indyk. High Dimensional Computational Geometry. PhD. Thesis, 2000.
[14] Piotr Indyk. On approximate nearest neighbors under l? norm. J. Comput. Syst. sci., 63(4),
2001.
[15] W. Johnson and J. Lindenstrauss. Extensions of lipschitz maps into a hilbert space. Contemp.
Math., 26:189?206, 1984.
[16] Norio Katayama and Shin?ichi Satoh. The SR-tree: an index structure for high-dimensional
nearest neighbor queries. pages 369?380, 1997.
[17] J. Kleinberg. Two Algorithms for Nearest Neighbor Search in High Dimension. In Proceedings
of the Twenty-ninth Annual ACM Symposium on the Theory of Computing, pages 599?608,
1997.
[18] E. Kushilevitz, R. Ostrovsky, and Y. Rabani. Efficient Search for Approximate Nearest Neighbors in High Dimensional Spaces. In Proceedings of the Thirtieth Annual ACM Symposium on
the Theory of Computing, 1998.
[19] T. Liu, A. W. Moore, A. Gray, and Ke. Yang. An investigation of practical approximate nearest
neighbor algorithms (full version). Manuscript in preparation.
[20] B. S. Manjunath. Airphoto dataset, http://vivaldi.ece.ucsb.edu/Manjunath/research.htm.
[21] B. S. Manjunath and W. Y. Ma. Texture features for browsing and retrieval of large image data.
IEEE Transactions on Pattern Analysis and Machine Intelligence, 18(8):837?842, 1996.
[22] A. W. Moore. The Anchors Hierarchy: Using the Triangle Inequality to Survive HighDimensional Data. In Twelfth Conference on Uncertainty in Artificial Intelligence. AAAI Press,
2000.
[23] G. Mori, S. Belongie, and J. Malik. Shape contexts enable efficient retrieval of similar shapes.
In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2001.
[24] S. M. Omohundro. Efficient Algorithms with Neural Network Behaviour. Journal of Complex
Systems, 1(2):273?347, 1987.
[25] S. M. Omohundro. Bumptrees for Efficient Function, Constraint, and Classification Learning.
In R. P. Lippmann, J. E. Moody, and D. S. Touretzky, editors, Advances in Neural Information
Processing Systems 3. Morgan Kaufmann, 1991.
[26] F. P. Preparata and M. Shamos. Computational Geometry. Springer-Verlag, 1985.
[27] Y. Rubnet, C. Tomasi, and L. J. Guibas. The earth mover?s distance as a metric for image
retrieval. International Journal of Computer Vision, 40(2):99?121, 2000.
[28] Gregory Shakhnarovich, Paul Viola, and Trevor Darrell. Fast pose estimation with parameter
sensitive hashing. In Proceedings of the International Conference on Computer Vision, 2003.
[29] J. K. Uhlmann. Satisfying general proximity/similarity queries with metric trees. Information
Processing Letters, 40:175?179, 1991.
[30] P. Yianilos. Excluded middle vantage point forests for nearest neighbor search. In DIMACS
Implementation Challenge, 1999.
| 2666 |@word repository:1 version:5 briefly:1 middle:1 norm:4 duda:1 nd:2 disk:8 twelfth:1 vldb:3 q1:1 pick:2 liu:2 series:2 contains:2 ours:4 outperforms:3 readily:1 john:1 partition:4 kdd:1 shape:2 designed:2 hash:4 v:2 intelligence:2 leaf:3 selected:1 unacceptably:1 plane:6 desktop:1 core:1 lr:7 hypersphere:2 certificate:1 math:1 node:31 org:1 simpler:1 five:1 rc:14 along:3 mathematical:1 become:1 symposium:3 descendant:1 prove:1 manner:1 introduce:4 expected:1 roughly:2 p1:6 nor:1 examine:1 frequently:1 actual:1 curse:4 little:2 cpu:9 becomes:7 project:3 revision:1 lowest:1 defeatist:10 kind:1 spends:2 finding:2 guarantee:1 sky:1 every:1 runtime:1 wrong:1 ostrovsky:1 partitioning:4 control:1 farthest:2 appear:2 before:1 negligible:1 mistake:1 severely:1 mount:1 bumptrees:1 twice:1 conversely:1 fastest:2 limited:2 range:3 practical:4 practice:6 block:1 implement:1 differs:1 silverman:1 procedure:2 shin:1 area:1 empirical:2 significantly:1 projection:8 word:2 pre:2 regular:1 vantage:1 get:1 onto:1 close:4 put:1 context:2 www:1 conventional:3 map:1 quick:1 center:3 demonstrated:1 go:3 attention:1 starting:1 survey:1 ke:2 simplicity:2 splitting:3 autonlab:1 pure:1 constrast:1 rule:1 kushilevitz:1 datapoints:6 embedding:1 searching:10 coordinate:1 construction:4 suppose:1 hierarchy:1 user:1 exact:8 pa:1 recognition:2 particularly:1 satisfying:1 database:3 solved:1 verifying:1 region:2 complexity:1 ideally:1 dynamic:1 shakhnarovich:1 efficiency:5 completely:1 triangle:1 easily:1 htm:1 stock:1 represented:4 various:1 separated:1 fast:2 effective:2 describe:1 query:12 artificial:1 kevin:1 shamos:1 apparent:1 whose:2 posed:1 heuristic:1 solve:2 say:2 distortion:1 otherwise:1 quite:1 corelfeatures:2 prh:2 indyk:6 reproduced:1 obviously:2 advantage:1 lowdimensional:1 uci:6 aligned:1 achieve:2 forth:1 description:1 kv:1 pronounced:1 parent:1 motwani:2 darrell:1 r1:3 produce:1 help:1 depending:1 andrew:1 pose:1 nearest:27 school:1 received:1 p2:6 descends:1 c:1 skip:1 implies:1 guided:1 radius:2 dfs:17 opteron:1 awm:1 centered:2 enable:1 public:1 behaviour:1 assign:1 fix:1 investigation:2 mathematically:1 frontier:1 strictly:1 extension:1 hold:2 proximity:2 ic:1 guibas:1 earth:1 estimation:1 proc:2 lose:1 uhlmann:1 sensitive:7 largest:2 tool:2 always:1 aim:1 rather:3 beyer:1 avoid:1 finkel:1 varying:1 thirtieth:1 publication:1 conjunction:1 improvement:1 consistently:2 rank:1 greatly:1 contrast:1 sense:1 voronoi:1 nn:26 inaccurate:1 typically:4 entire:1 sb:1 pixel:1 overall:1 dual:1 among:1 html:1 denoted:1 classification:2 spatial:4 art:1 having:1 piotr:2 eliminated:1 sampling:1 identical:1 represents:2 survive:1 report:1 preparata:1 few:1 randomly:2 ve:6 mover:1 geometry:2 maintain:1 friedman:1 highly:2 evaluation:1 accurate:3 daily:1 orthogonal:1 traversing:1 tree:101 modest:1 euclidean:1 machinery:1 earlier:1 cost:1 introducing:1 subset:3 entry:1 successful:2 examining:1 conducted:2 johnson:4 answer:2 varies:4 gregory:1 considerably:1 international:3 standing:1 guttman:1 quickly:1 moody:1 na:6 linux:1 thesis:1 aaai:1 containing:1 choose:2 ucsb:1 resort:1 inefficient:2 return:1 syst:1 nonoverlapping:1 summarized:2 suse:1 gionis:1 notable:1 sloan:1 later:1 root:1 doing:4 start:1 competitive:1 maintains:1 capability:1 parallel:1 publicly:1 accuracy:7 qk:1 kaufmann:1 basically:1 drive:1 researcher:1 processor:1 explain:1 touretzky:1 whenever:1 trevor:1 lengthy:1 failure:1 nonetheless:2 galaxy:6 naturally:1 stop:1 lpv:12 dataset:9 proved:1 popular:1 ask:1 color:2 realm:1 dimensionality:6 hilbert:1 actually:1 back:2 goldstein:3 manuscript:1 hashing:7 higher:1 april:1 done:1 though:1 furthermore:2 just:1 until:1 hand:2 overlapping:14 perhaps:2 gray:2 believe:1 bentley:1 usa:1 building:1 contain:4 true:5 verify:1 excluded:1 moore:3 deal:1 round:4 ll:6 during:2 criterion:1 dimacs:1 omohundro:2 performs:1 l1:2 rpv:11 image:5 wise:1 ranging:1 discovers:1 mt:15 enthusiasm:1 corel:14 belong:5 mellon:1 significant:1 rd:3 satoh:2 lsh:21 access:3 similarity:3 closest:2 own:3 recent:2 moderate:1 apart:1 certain:1 buffer:5 verlag:1 inequality:1 binary:1 arbitrarily:1 continue:1 morgan:1 greater:3 relaxed:1 prune:1 kilo:1 patella:1 multiple:1 full:2 reduces:1 match:1 cross:1 long:1 sphere:1 retrieval:3 hart:1 equally:1 visit:1 parenthesis:1 vivaldi:1 variant:3 vision:6 cmu:1 metric:43 histogram:2 represent:2 separately:1 diagram:1 median:2 rest:1 sr:14 sigact:1 undo:1 member:1 seem:1 contemp:1 call:5 yang:2 split:4 enough:1 variety:1 fit:1 perfectly:1 opposite:1 reduce:1 idea:2 pivot:3 motivated:1 gb:1 manjunath:3 clarkson:1 suffer:2 returned:1 speaking:2 useful:2 detailed:1 clear:1 mid:2 http:3 assn:1 exist:1 notice:4 disjoint:2 per:1 thumbnail:1 carnegie:1 write:1 ichi:1 key:1 thereafter:1 threshold:3 achieving:2 ram:1 fraction:3 year:1 sum:1 realworld:1 run:5 letter:1 powerful:1 uncertainty:1 place:1 throughout:1 family:2 reasonable:1 decide:2 strange:1 almost:2 wu:1 decision:10 comparable:4 fold:4 annual:2 adapted:1 ahead:1 precisely:1 constraint:1 scene:1 software:1 autonweb:1 kleinberg:1 speed:6 rabani:1 extremely:1 performing:1 speedup:9 ball:3 aerial:10 kd:2 across:1 remain:1 increasingly:1 son:1 partitioned:1 modification:2 happens:1 making:1 bucket:3 taken:2 mori:1 zezula:1 remains:2 turn:1 mechanism:1 needed:1 photo:1 raghu:1 available:3 operation:2 permit:1 observe:2 hierarchical:1 away:1 appropriate:1 slower:1 original:2 denotes:1 running:1 ensure:2 spill:9 somewhere:1 exploit:2 sigmod:1 ting:1 especially:1 malik:1 question:1 ciaccia:1 strategy:1 traditional:3 september:2 kth:1 subspace:4 distance:17 unable:2 separating:1 sci:1 amd:1 extent:1 besides:1 code:1 index:2 ratio:1 balance:3 stoc:1 trace:7 stated:1 implementation:9 twenty:1 perform:4 redesign:1 observation:1 datasets:14 arya:1 benchmark:1 defining:1 viola:1 ninth:1 namely:3 pair:1 optimized:1 tomasi:1 able:2 beyond:1 usually:2 perception:1 pattern:4 challenge:1 max:1 memory:2 overlap:2 hybrid:12 representing:1 scheme:1 altered:1 library:2 naive:1 katayama:2 byte:1 review:1 l2:1 expect:2 lecture:1 querying:1 digital:1 validation:1 verification:1 consistent:1 principle:1 editor:1 intractability:1 netanyahu:1 summary:1 repeat:1 surprisingly:1 enjoys:1 side:2 allow:2 telling:1 neighbor:26 wide:1 fall:2 ghz:1 boundary:9 dimension:19 depth:7 world:2 lindenstrauss:4 author:2 commonly:1 projected:2 far:4 transaction:2 approximate:22 lippmann:1 dealing:1 hist:8 anchor:1 pittsburgh:1 belongie:1 spectrum:3 search:60 why:1 table:4 nature:1 terminate:1 forest:1 agray:1 necessarily:1 complex:1 yianilos:2 sp:30 whole:1 paul:1 n2:1 child:12 slow:1 wiley:1 lc:15 shaft:1 sub:1 fails:1 comput:1 candidate:3 third:2 dozen:1 down:3 removing:1 embed:1 specific:1 r2:3 explored:1 concern:1 evidence:1 exists:1 intrinsic:1 texture:3 magnitude:2 phd:1 uri:1 browsing:1 locality:4 logarithmic:2 backtracking:4 simply:2 explore:1 likely:1 photograph:1 failed:1 springer:1 ramakrishnan:3 corresponds:1 acm:5 ma:1 marked:1 viewed:1 sized:1 acceleration:1 towards:1 shared:2 lipschitz:1 considerable:1 change:1 content:1 typical:1 specifically:1 except:1 lemma:3 called:1 total:1 ece:1 experimental:3 meaningful:1 organizes:1 formally:2 highdimensional:1 internal:2 support:1 searched:4 mark:1 scan:2 latter:2 alexander:1 jonathan:1 preparation:1 |
1,835 | 2,667 | Using Random Forests in the Structured
Language Model
Peng Xu and Frederick Jelinek
Center for Language and Speech Processing
Department of Electrical and Computer Engineering
The Johns Hopkins University
{xp,jelinek}@jhu.edu
Abstract
In this paper, we explore the use of Random Forests (RFs) in the structured language model (SLM), which uses rich syntactic information in
predicting the next word based on words already seen. The goal in this
work is to construct RFs by randomly growing Decision Trees (DTs) using syntactic information and investigate the performance of the SLM
modeled by the RFs in automatic speech recognition.
RFs, which were originally developed as classifiers, are a combination
of decision tree classifiers. Each tree is grown based on random training
data sampled independently and with the same distribution for all trees in
the forest, and a random selection of possible questions at each node of
the decision tree. Our approach extends the original idea of RFs to deal
with the data sparseness problem encountered in language modeling.
RFs have been studied in the context of n-gram language modeling and
have been shown to generalize well to unseen data. We show in this paper
that RFs using syntactic information can also achieve better performance
in both perplexity (PPL) and word error rate (WER) in a large vocabulary
speech recognition system, compared to a baseline that uses Kneser-Ney
smoothing.
1
Introduction
In many systems dealing with speech or natural language, such as Automatic Speech
Recognition and Statistical Machine Translation, a language model is a crucial component
for searching in the often prohibitively large hypothesis space. Most state-of-the-art systems use n-gram language models, which are simple and effective most of the time. Many
smoothing techniques that improve language model probability estimation have been proposed and studied in the n-gram literature [1]. There has so far been work in exploring
Decision Tree (DT) language models [2, 3], which attempt to cluster similar histories together to achieve better probability estimation. However, the results were negative [3]:
decision tree language models failed to improve upon the baseline n-gram models with the
same order n.
Random Forest (RF) language models, which are generalizations of DT language models,
have been recently applied to word n-grams [4]. DT growing is randomized to construct
RFs efficiently. Once constructed, the RFs function as a randomized history clustering,
which helps in dealing with the data sparseness problem. In general, the weakness of some
trees can be compensated for by other trees. The collective contribution of all DTs in an
RF n-gram model results in significant improvements in both perplexity (PPL) and word
error rate (WER) in a large vocabulary speech recognition system.
Language models can also be improved with better representations of the history. Recent
efforts have studied various ways of using information from a longer history span than that
usually captured by normal n-gram language models, as well as ways of using syntactic
information that is not available to the word-based n-gram models [5, 6, 7]. All these language models are based on stochastic parsing techniques that build up parse trees for the
input word sequence and condition the generation of words on syntactic and lexical information available in the parse trees. Since these language models capture useful hierarchical
characteristics of language, they can improve PPL and WER significantly for various tasks.
However, due to the n-gram nature of the components of the syntactic language models,
the data sparseness problem can be severe.
In order to reduce the data sparseness problem for using rich syntactic information in the
context, we study the use of RFs in the structured language model (SLM) [5]. Our results
show that although the components of the SLM have high order n-grams, our RF approach
can still achieve better performance, reducing both the perplexity (PPL) and word error
rate (WER) in a large vocabulary speech recognition system compared to a Kneser-Ney
smoothing baseline.
2
Basic Language Modeling
The purpose of a language model is to estimate the probability of a word string. Let W
denote a string of N words, that is, W = w1 , w2 , . . . , wN . Then, by the chain rule of
probability, we have
Q
P (W )=P (w1 )? N
(1)
i=2 P (wi |w1 ,...,wi?1 ).
In order to estimate the probabilities P (wi |w1 , . . . , wi?1 ), we need a training corpus consisting of a large number of words. However, in any practical natural language system of
even moderate vocabulary size, it is clear that the number of probabilities to be estimated
and stored is prohibitively large. Therefore, histories w1 , . . . , wi?1 for a word wi are usually grouped into equivalence classes. The most widely used language models, n-gram
language models, use the identities of the last n ? 1 words as equivalence classes. In an
n-gram model, we then have
P (W )=P (w1 )?
QN
i=2
i?1
P (wi |wi?n+1
),
(2)
i?1
where we have used wi?n+1
to denote the word sequence wi?n+1 , . . . , wi?1 .
If we could handle unlimited amounts of training data, the maximum likelihood (ML)
i?1
estimate of P (wi |wi?n+1
) would be the best:
i?1
P (wi |wi?n+1
)=
i
C(wi?n+1
)
,
i?1
C(w
)
i?n+1
(3)
i
where C(wi?n+1
) is the number of times the string wi?n+1 , . . . , wi is seen in the training
data.
2.1
Language Model Smoothing
An n-gram model when n = 3 is called a trigram model. For a vocabulary of size |V | =
104 , there are |V |3 = 1012 trigram probabilities to be estimated. For any training data of a
manageable size, many of the probabilities will be zero if the ML estimate is used.
In order to solve this problem, many smoothing techniques have been studied (see [1]
and the references therein). Smoothing adjusts the ML estimates to produce more accurate
probabilities and to assign nonzero probabilities to any word string. Details about various
smoothing techniques will not be presented in this paper, but we will outline a particular
way of smoothing, namely interpolated Kneser-Ney smoothing [8], for later reference.
Interpolated Kneser-Ney smoothing assumes the following form:
i?1
PKN (wi |wi?n+1
)
=
i
max(C(wi?n+1
)?D,0)
i?1
C(w
)
i?n+1
(4)
i?1
i?1
+?(wi?n+1
)PKN (wi |wi?n+2
),
i?1
where D is a discounting constant and ?(wi?n+1
) is the interpolation weight for the lower
order probabilities ((n ? 1)-gram). The discount constant is often estimated using the
n1
leave-one-out method, leading to the approximation D = n1 +2n
, where n1 is the number
2
of n-grams with count one and n2 is the number of n-grams with count two. To ensure that
the probabilities sum to one, we have
D
P
i?1
?(wi?n+1
)=
wi :C(wi
)>0
i?n+1
i?1
C(w
)
i?n+1
1
.
The lower order probabilities in interpolated Kneser-Ney smoothing can be estimated as
(assuming ML estimation):
P
i?1
PKN (wi |wi?n+2
)= P
1
wi?n+1 :C(wi
)>0
i?n+1
.
1
wi?n+1 ,wi :C(wi
)>0
i?n+1
(5)
Note that the lower order probabilities are usually recursively smoothed using Equation 4.
2.2
Language Model Evalution
A commonly used task-independent quality measure for a given language model is related
to the cross-entropy of the underlying model and is referred to as perplexity (PPL):
P P L=exp(?1/N
PN
i=1
log [P (wi |w1i?1 )]),
(6)
where w1 , . . . , wN is the test text that consists of N words.
For different tasks, there are different task-dependent quality measures of language models. For example, in an automatic speech recognition system, the performance is usually
measured by word error rate (WER).
3
The Structured Language Model (SLM)
The SLM uses rich syntactic information beyond regular word n-grams to improve language model quality. An extensive presentation of the SLM can be found in Chelba and
Jelinek, 2000 [5]. The model assigns a probability P (W, T ) to every sentence W and every possible binary parse T . The terminals of T are the words of W with POS tags, and
the nodes of T are annotated with phrase headwords and non-terminal labels. Let W be
h_{-m} = (<s>, SB)
h_{-1}
h_0 = (h_0.word, h_0.tag)
(<s>, SB) ....... (w_p, t_p) (w_{p+1}, t_{p+1}) ........ (w_k, t_k) w_{k+1}.... </s>
Figure 1: A word-parse k-prefix
a sentence of length n words to which we have prepended the sentence beginning marker
<s> and appended the sentence end marker </s> so that w0 =<s> and wn+1 =</s>.
Let Wk = w0 . . . wk be the word k-prefix of the sentence ? the words from the beginning
of the sentence up to the current position k ? and Wk Tk the word-parse k-prefix. Figure 1
shows a word-parse k-prefix; h_0, .., h_{-m} are the exposed heads, each head being a pair (headword, non-terminal label), or (word, POS tag) in the case of a root-only
tree. The exposed heads at a given position k in the input sentence are a function of the
word-parse k-prefix [5].
The joint probability P (W, T ) of a word sequence W and a complete parse T comes from
contributions of three components: WORD-PREDICTOR, TAGGER and CONSTRUCTOR. The SLM works in the following way: first, the WORD-PREDICTOR predicts a
word based on the word-parse prefix; the TAGGER then assigns a POS tag to the predicted
word based on the word itself and the word-parse prefix; the CONSTRUCTOR takes a
series of actions each of which turns a parse prefix into a new parse prefix (the series of
actions ends with a NULL action which tells the WORD-PREDICTOR to predict the next
word). Details about the three components can be found in [5]. Each of the three components can be seen as an n-gram model and can be estimated independently because of the
product form of the joint probability. They are parameterized (approximated) as follows:
P (wk |Wk?1 Tk?1 )
=
P (wk |h0 .tag,h0 .word,h?1 .tag,h?1 .word),
P (tk |wk ,Wk?1 Tk?1 )
=
P (tk |wk ,h0 .tag,h?1 .tag),
k
k
P (pk
i |Wk?1 Tk?1 ,wk ,tk ,p1 ...pi?1 )
=
P (pk
i |h0 .tag,h?1 .tag,h?2 .tag,h0 .word,h?1 .word),
(7)
(8)
(9)
where pki is the ith CONSTRUCTOR action after the k th word and POS tag have been
predicted. Since the number of parses for a given word prefix Wk grows exponentially
with k, |{Tk }| ? O(2k ), the state space of our model is huge even for relatively short
sentences. Thus we must use a search strategy that prunes the allowable parse set. One
choice is a synchronous multi-stack search algorithm [5] which is very similar to a beam
search.
The language model probability assignment for the word at position k + 1 in the input
sentence is made using:
PSLM (wk+1 |Wk )
=
P
?(Wk ,Tk )
=
P (Wk Tk )/
Tk ?Sk
P (wk+1 |Wk Tk )??(Wk ,Tk ),
P
T ?S P (Wk Tk ),
k
k
(10)
which ensures a proper probability normalization over strings of words, where S k is the set
of all parses present in the stacks at the current stage k and P (Wk Tk ) is the joint probability
of word-parse prefix Wk Tk .
Each model component ?WORD-PREDICTOR, TAGGER, CONSTRUCTOR? is estimated independently from a set of parsed sentences after undergoing headword percolation
and binarization (see details in [5]).
4
4.1
Using Random Forests in the Structured Language Model
Random Forest n-gram Modeling
A Random Forest (RF) n-gram model is a collection of randomly constructed decision tree
(DT) n-gram models. Unlike RFs in classification and regression tasks [9, 10, 11], RFs
are used in language modeling to deal with the data sparsenes problem [4]. Therefore, the
training data is not randomly sampled for each DT. Figure 2 shows the algorithm DT-Grow
and Node-Split used for generating random DT language models.
We define a position in the history as the distance between a word in the history and the
predicted word. The randomization is carried out in two places: a random selection of
Algorithm DT-Grow
Input: counts for training and heldout data
Algorithm Node-Split(p)
Input: node p and training data associated
Create a root node containing all
Initialize: histories in the training data and
put it in set ?
Initialize:
While ? is not empty
1. Get a node p from ?
2. If Node-Split(p) is successful, eliminate p from ? and put the two children of p in ?
Foreach internal node p in the tree
Randomly select a subset of positions I in the history
Foreach position i in I
1. Group all histories into basic elements
?(v)
2. Randomly split the elements ?(v)
into sets L and R
3. While there are elements moved, Do
1. LH
p ? normalized likelihood of heldout data associated with p, using
training data statistics in p
(a) Move each element from L to
R if the move results in positive
gain in training data likelihood
2. Get the set of leaves P rooted in p
3. LH
P ? normalized likelihood of heldout data associated with all leaves in
P, using training data statistics in the
corresponding leaves
H
4. if LH
P ? Lp < 0, prune the subtree
rooted in p
Output: a Decistion Tree language model
(b) Move each element from R to
L if the move results in positive
gain in training data likelihood
Select the position from I that results in the
largest gain
Output: a split L and R, or failure if the largest
gain is not positive
Figure 2: The algorithm DT-Grow and Node-Split
positions in the history and an initial random split of basic elements. Since our splitting
criterion is to maximize the log-likelihood of the training data, each split uses only statistics
(from training data) associated with the node under consideration. Smoothing is not needed
in the splitting and we can use a fast exchange algorithm [12] in Node-Split. Given a
position i in the history, ?(v) is defined to be the set of histories belonging to the node p,
such that they all have word v at position i. It is clear that for every position i in the history,
the union ?v ?(v) is all histories in the node p.
In DT-Grow, after a DT is fully grown, we use some heldout data to prune it. Pruning is
done in such a way that we maximize the likelihood of the heldout data, where smoothing
is applied according to Equation 4:
i?1
PDT (wi |?DT (wi?n+1
))
=
i?1
))?D,0)
i?n+1
i?1
C(?DT (w
))
i?n+1
max(C(wi ,?DT (w
(11)
i?1
i?1
+?(?DT (wi?n+1
))PKN (wi |wi?n+2
)
i?1
where ?DT (?) is one of the DT nodes the history can be mapped to and PKN (wi |wi?n+2
)
is as defined in Equation 5. This pruning is similar to the pruning strategy used in
CART [13].
Once we get the DTs, we only use the leaf nodes as equivalence classes of histories. If
a new history is encountered, it is very likely that we will not be able to place it at a leaf
i?1
node in the DT. In this case, ?(?DT (wi?n+1
)) = 1 in Equation 11 and we simply use
i?1
PKN (wi |wi?n+2 ) to get the probabilities.
The randomized version of the DT growing algorithm can be run many times and finally
we will get a collection of randomly grown DTs: a Random Forest (RF). Since each DT is
a smoothed language model, we simply aggregate all DTs in our RF to get the RF language
model. Suppose we have M randomly grown DTs, DT1 , . . . , DTM . In the n-gram case,
the RF language model probabilities can be computed as:
i?1
1
PRF (wi |wi?n+1
)= M
i?1
?DTj (wi?n+1
)
PM
j=1
(12)
i?1
PDTj (wi |?DTj (wi?n+1
))
i?1
wi?n+1
i?1
wi?n+1
where
maps the history
to a leaf node in DTj . If
can not
be mapped to a leaf node in some DT, we back-off to the lower order KN probability as
mentioned earlier.
It can be shown by the Law of Large Numbers that the probability in Equation
12 converges
i?1
as the number of DTs grows. It converges to ET PT (wi |?T (wi?n+1
) where T is a
random variable representing the random DTs. The advantage of the RF approach over the
KN smoothing lies in the fact that different DTs have different weaknesses and strengths
for word prediction. As the number of trees grows, the weakness of some trees can be
compensated for by some other trees. This advantage and the convergence have been shown
experimentally in [4].
4.2
Using RFs in the SLM
Since the three model components in the SLM as in Equation 7-9 can be estimated independently, we can construct an RF for each component using the algorithm DT-Grow in
the previous section. The only difference is that we will have different n-gram orders and
different items in the history for each model.
Ideally, we would like to use RFs for each component in the SLM. However, due to the
nature of the SLM, there are difficulties. The SLM uses a synchronous multi-stack search
algorithm to dynamically construct stacks and compute the language model probabilities
as in Equation 10. If we use RFs for all components, we need to load all DTs in the RFs
into memory at runtime. This is impractical for RFs of any reasonable size.
There is a different approach that can take advantage of the randomness in the RFs. Supa
pose we have M randomly grown DTs, DT1a , . . . , DTM
for each component a of the SLM,
where a ? {P, T, C} for WORD-PREDICTOR, TAGGER and CONSTRUCTOR, respectively. The DTs are grouped into M triples {DTjP , DTjT , DTjC } j = 1, . . . , M . We
calculate the joint probability P (W, T ) for the j th DT triple according to:
Pj (W,T )=
Qn+1
k=1 [PDT P
j
(wk |Wk?1 Tk?1 )?PDT T (tk |Wk?1 Tk?1 ,wk )?
j
Q Nk
k
k
k
i=1 PDT C (pi |Wk?1 Tk?1 ,wk ,tk ,p1 ...pi?1 )].
j
(13)
Then, the language model probability assignment for the j th DT triple is made using:
Pj (wk+1 |Wk )
=
?j (Wk ,Tkj )
=
j
j
j PDT P (wk+1 |Wk Tk )??j (Wk ,Tk ),
j
T ?Sk
j
k
P
j
Pj (Wk Tkj )/
j
j Pj (Wk Tk ),
T ?S
k
k
P
(14)
which is achieved by running the synchronous multi-stack algorithm using the j th DT triple
as a model. Finally, after the SLM is run M times, the RF language model probability is
an average of the probabilities above:
1
PRF (wk+1 |Wk )= M
PM
j=1
Pj (wk+1 |Wk ).
(15)
The triple {DTjP , DTjT , DTjC } can be considered as a single DT in which the root node
has three children corresponding to the three root nodes of DTjP , DTjT and DTjC . The
root node of this DT asks the question: Which model component does the history belong
to? According to the answer, we can proceed to one of the three children nodes (one of the
three components, in fact). Since the multi-stack search algorithm is deterministic given
the DT, the probability in Equation 15 can be shown to converge.
5
5.1
Experiments
Perplexity (PPL)
We have used the UPenn Treebank portion of the WSJ corpus to carry out our experiments.
The UPenn Treebank contains 24 sections of hand-parsed sentences, for a total of about
one million words. We used section 00-20 for training our models, section 21-22 as heldout
data for pruning the DTs, and section 23-24 to test our models. Before carrying out our
experiments, we normalized the text in the following ways: numbers in arabic form were
replaced by a single token ?N?, punctuation was removed, all words were mapped to lower
case, extra information in the parse trees was ignored, and, finally, traces were ignored. The
word vocabulary contains 10k words including a special token for unknown words. There
are 40 items in the part-of-speech set and 54 items in the non-terminal set, respectively.
The three components in the SLM were treated independently during training. We trained
an RF for each component and each RF contained 100 randomly grown DTs. The baseline
SLM used KN smoothing (KN-SLM). The 100 probability sequences from the 100 triples
were aggregated to get the final PPL. The results are shown in Table 1. We also interpolated
the SLM with the KN-trigram to get further improvements. The interpolation weight ? in
Table 1 is on KN-trigram. The RF-SLM achieved a 10.9% and a 7.5% improvement over
the KN-SLM, before and after interpolation with KN-trigram, respectively. Compared to
the improvements reported in [4] (10.5% from RF-trigram to KN-trigram), the RF-SLM
achieved greater improvement by using syntactic information. Figure 3 shows the convergence of the PPL as the number of DTs grows from 1 to 100.
160
155
Model
KN-SLM
RF-SLM
Gain
150
PPL
145
140
135
?=0.0
137.9
122.8
10.9%
?=0.4
127.2
117.6
7.5%
?=1.0
145.0
145.0
-
130
Table 1: PPL comparison between KNSLM and RF-SLM, interpolated with
KN-trigram
125
120
20
40
60
Number of DTs
80
100
Figure 3: PPL convergence
5.2
Word Error Rate by N -best Re-scoring
To test our RF modeling approach in the context of speech recognition, we evaluated the
models in the WSJ DARPA?93 HUB1 test setup. The size of the test set is 213 utterances,
3446 words. The 20k word open vocabulary and baseline 3-gram model are the standard
ones provided by NIST and LDC ? see [5] for details. The N -best lists were generated
using the standard 3-gram model trained on 40M words of WSJ. The N -best size was at
most 50 for each utterance, and the average size was about 23. For the KN-SLM and RFSLM, we used 20M words automatically parsed, binarized and enriched with headwords
and NT/POS tag information. As the size of RF-SLM becomes very large, we only used
RF for the WORD-PREDICTOR component (RF-SLM-P). The other two components used
KN smoothing. The results are reported in Table 2.
Model
KN-SLM
RF-SLM-P
?=0.0
12.8
11.9
?=0.2
12.5
12.2
?=0.4
12.6
12.3
?=0.6
12.7
12.3
?=0.8
12.7
12.6
Table 2: N-best rescoring WER results
For purpose of comparison, we interpolated all models with the KN-trigram built from
40M words at different level of interpolation weights ? (on KN-trigram). However, it is the
? = 0.0 column that is the most interesting. We can see that the RF approach improved
over the regular KN approach with an absolute WER reduction of 0.9%.
6
Conclusions
Based on the idea of Random Forests in classification and regression, we developed algorithms for constructing and using Random Forests in language modeling. In particular, we
applied this new probability estimation technique to the Structured Language Model, in
which there are three model components that can be estimated independently. The independently constructed Random Forests can be considered as a more general single Random
Forest, which ensures the convergence of the probabilities as the number of Decision Trees
grows. The results on a large vocabulary speech recognition system show that we can
achieve significant reduction in both perplexity and word error rate, compared to a baseline
using Kneser-Ney smoothing.
References
[1] Stanley F. Chen and Joshua Goodman, ?An empirical study of smoothing techniques for language modeling,? Tech. Rep. TR-10-98, Computer Science Group, Harvard University, Cambridge, Massachusetts, 1998.
[2] L. Bahl, P. Brown, P. de Souza, and R. Mercer, ?A tree-based statistical language model for
natural language speech recognition,? in IEEE Transactions on Acoustics, Speech and Signal
Processing, July 1989, vol. 37, pp. 1001?1008.
[3] Gerasimos Potamianos and Frederick Jelinek, ?A study of n-gram and decision tree letter
language modeling methods,? Speech Communication, vol. 24(3), pp. 171?192, 1998.
[4] Peng Xu and Frederick Jelinek, ?Random forests in language modeling,? in Proceedings of the
2004 Conference on Empirical Methods in Natural Language Processing, Barcelona, Spain,
July 2004.
[5] Ciprian Chelba and Frederick Jelinek, ?Structured language modeling,? Computer Speech and
Language, vol. 14, no. 4, pp. 283?332, October 2000.
[6] Eugene Charniak, ?Immediate-head parsing for language models,? in Proceedings of the 39th
Annual Meeting and 10th Conference of the European Chapter of ACL, Toulouse, France, July
2001, pp. 116?123.
[7] Brian Roark, Robust Probabilistic Predictive Syntactic Processing: Motivations, Models and
Applications, Ph.D. thesis, Brown University, Providence, RI, 2001.
[8] Reinhard Kneser and Hermann Ney, ?Improved backing-off for m-gram language modeling,?
in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing, 1995, vol. 1, pp. 181?184.
[9] Y. Amit and D. Geman, ?Shape quantization and recognition with randomized trees,? Neural
Computation, , no. 9, pp. 1545?1588, 1997.
[10] Leo Breiman, ?Random forests,? Tech. Rep., Statistics Department, University of California,
Berkeley, Berkeley, CA, 2001.
[11] T.K. Ho, ?The random subspace method for constructing decision forests,? IEEE Trans. on
Pattern Analysis and Machine Intelligence, vol. 20, no. 8, pp. 832?844, 1998.
[12] S. Martin, J. Liermann, and H. Ney, ?Algorithms for bigram and trigram word clustering,?
Speech Communication, vol. 24(3), pp. 171?192, 1998.
[13] L. Breiman, J.H. Friedman, R.A. Olshen, and C.J. Stone, Classification and Regression Trees,
Chapman and Hall, New York, 1984.
| 2667 |@word arabic:1 version:1 manageable:1 bigram:1 open:1 t_:1 asks:1 tr:1 recursively:1 carry:1 reduction:2 initial:1 series:2 contains:2 charniak:1 prefix:11 current:2 nt:1 must:1 parsing:2 john:1 shape:1 headword:4 intelligence:1 leaf:7 item:3 beginning:2 prepended:1 ith:1 short:1 rescoring:1 node:23 tagger:4 constructed:3 consists:1 peng:2 upenn:2 p1:2 chelba:2 growing:3 multi:4 terminal:4 automatically:1 becomes:1 provided:1 spain:1 underlying:1 null:1 string:5 developed:2 impractical:1 berkeley:2 every:3 binarized:1 runtime:1 prohibitively:2 slm:30 classifier:2 positive:3 before:2 engineering:1 interpolation:4 kneser:7 acl:1 therein:1 studied:4 equivalence:3 dynamically:1 practical:1 union:1 empirical:2 jhu:1 significantly:1 word:68 regular:2 gerasimos:1 get:8 selection:2 put:2 context:3 map:1 deterministic:1 center:1 compensated:2 lexical:1 independently:7 splitting:2 assigns:2 rule:1 adjusts:1 searching:1 handle:1 constructor:5 pt:1 suppose:1 us:5 hypothesis:1 harvard:1 element:6 recognition:10 approximated:1 predicts:1 geman:1 electrical:1 capture:1 calculate:1 ensures:2 tkj:2 removed:1 mentioned:1 ideally:1 trained:2 carrying:1 exposed:2 predictive:1 upon:1 po:5 joint:4 darpa:1 various:3 chapter:1 grown:6 leo:1 fast:1 effective:1 tell:1 aggregate:1 h0:5 widely:1 solve:1 toulouse:1 statistic:4 unseen:1 syntactic:10 itself:1 final:1 sequence:4 advantage:3 product:1 achieve:4 moved:1 convergence:4 cluster:1 empty:1 produce:1 generating:1 wsj:3 leave:1 converges:2 tk:24 help:1 pose:1 measured:1 predicted:3 come:1 hermann:1 annotated:1 stochastic:1 exchange:1 assign:1 generalization:1 randomization:1 brian:1 exploring:1 considered:2 hall:1 normal:1 exp:1 predict:1 pkn:6 trigram:11 purpose:2 estimation:4 label:2 percolation:1 prf:2 grouped:2 largest:2 create:1 pki:1 pn:1 breiman:2 improvement:5 likelihood:7 tech:2 baseline:6 dependent:1 sb:2 eliminate:1 france:1 backing:1 classification:3 smoothing:18 art:1 initialize:2 special:1 construct:4 once:2 ciprian:1 chapman:1 h_:3 randomly:9 replaced:1 consisting:1 n1:3 attempt:1 friedman:1 huge:1 investigate:1 severe:1 weakness:3 punctuation:1 chain:1 accurate:1 lh:3 tree:24 re:1 column:1 modeling:12 earlier:1 w1i:1 assignment:2 phrase:1 subset:1 predictor:6 successful:1 stored:1 reported:2 kn:17 answer:1 providence:1 international:1 randomized:4 probabilistic:1 off:2 together:1 hopkins:1 w1:7 thesis:1 containing:1 leading:1 de:1 w_:2 wk:40 later:1 root:5 portion:1 contribution:2 appended:1 characteristic:1 efficiently:1 generalize:1 randomness:1 history:21 failure:1 pp:8 associated:4 sampled:2 gain:5 massachusetts:1 stanley:1 back:1 originally:1 dt:29 improved:3 done:1 evaluated:1 stage:1 hand:1 parse:15 marker:2 bahl:1 quality:3 grows:5 normalized:3 brown:2 discounting:1 nonzero:1 deal:2 during:1 rooted:2 criterion:1 stone:1 allowable:1 outline:1 complete:1 consideration:1 recently:1 exponentially:1 foreach:2 million:1 belong:1 significant:2 cambridge:1 automatic:3 pm:2 language:54 longer:1 recent:1 moderate:1 perplexity:6 binary:1 rep:2 meeting:1 joshua:1 scoring:1 seen:3 captured:1 greater:1 prune:3 converge:1 maximize:2 aggregated:1 signal:2 july:3 cross:1 dtm:2 prediction:1 basic:3 regression:3 normalization:1 achieved:3 beam:1 grow:5 crucial:1 goodman:1 w2:1 extra:1 unlike:1 cart:1 split:9 wn:3 reduce:1 idea:2 synchronous:3 effort:1 speech:17 proceed:1 york:1 action:4 ignored:2 useful:1 clear:2 amount:1 discount:1 reinhard:1 ph:1 estimated:8 vol:6 ppl:11 group:2 pj:5 sum:1 run:2 parameterized:1 wer:7 letter:1 extends:1 place:2 reasonable:1 roark:1 decision:9 encountered:2 annual:1 strength:1 ri:1 unlimited:1 tag:13 interpolated:6 span:1 relatively:1 martin:1 structured:7 department:2 according:3 combination:1 belonging:1 wi:57 lp:1 potamianos:1 equation:8 turn:1 count:3 needed:1 end:2 available:2 hierarchical:1 pdt:5 ney:8 ho:1 original:1 assumes:1 clustering:2 ensure:1 running:1 parsed:3 build:1 amit:1 move:4 already:1 question:2 strategy:2 subspace:1 distance:1 mapped:3 w0:2 assuming:1 length:1 modeled:1 setup:1 october:1 olshen:1 trace:1 negative:1 collective:1 proper:1 unknown:1 nist:1 immediate:1 communication:2 head:4 supa:1 stack:6 smoothed:2 souza:1 namely:1 pair:1 extensive:1 sentence:11 t_p:1 acoustic:2 california:1 dts:16 barcelona:1 trans:1 frederick:4 beyond:1 able:1 usually:4 pattern:1 rf:42 max:2 memory:1 including:1 built:1 ldc:1 natural:4 difficulty:1 treated:1 predicting:1 representing:1 improve:4 carried:1 utterance:2 text:2 binarization:1 literature:1 eugene:1 law:1 fully:1 par:2 heldout:6 generation:1 interesting:1 triple:6 xp:1 mercer:1 treebank:2 pi:3 translation:1 token:2 last:1 absolute:1 jelinek:6 vocabulary:8 gram:27 rich:3 qn:2 commonly:1 made:2 collection:2 far:1 transaction:1 pruning:4 dealing:2 ml:4 corpus:2 search:5 sk:2 table:5 nature:2 robust:1 ca:1 forest:15 european:1 constructing:2 pk:2 motivation:1 n2:1 child:3 dtj:3 xu:2 enriched:1 referred:1 position:11 lie:1 load:1 undergoing:1 list:1 quantization:1 subtree:1 sparseness:4 nk:1 chen:1 entropy:1 simply:2 explore:1 likely:1 failed:1 contained:1 dt1:1 goal:1 identity:1 presentation:1 experimentally:1 reducing:1 called:1 total:1 select:2 internal:1 |
1,836 | 2,668 | A Hidden Markov Model for de Novo Peptide
Sequencing
Bernd Fischer, Volker Roth, Joachim M. Buhmann
Institute of Computational Science
ETH Zurich
CH-8092 Zurich, Switzerland
[email protected]
Jonas Grossmann, Sacha Baginsky,
Wilhelm Gruissem
Institute of Plant Sciences
ETH Zurich
CH-8092 Zurich, Switzerland
Franz Roos,
Peter Widmayer
Inst. of Theoretical Computer Science
ETH Zurich
CH-8092 Zurich, Switzerland
Abstract
De novo Sequencing of peptides is a challenging task in proteome research. While there exist reliable DNA-sequencing methods, the highthroughput de novo sequencing of proteins by mass spectrometry is still
an open problem. Current approaches suffer from a lack in precision
to detect mass peaks in the spectrograms. In this paper we present a
novel method for de novo peptide sequencing based on a hidden Markov
model. Experiments effectively demonstrate that this new method significantly outperforms standard approaches in matching quality.
1
Introduction
The goal of de novo peptide sequencing is to reconstruct an amino acid sequence from a
given mass spectrum. De novo sequencing by means of mass spectrometry is a very challenging task, since many practical problems like measurement errors or peak suppression
have to be overcome. It is, thus, not surprising that current approaches to reconstruct the
sequence from mass spectra are usually limited to those species for which genome information is available. This case is a simplified problem of the de novo sequencing problem,
since the hypothesis space of possible sequences is restricted to the known ones contained
in a sequence database.
In this paper we present a Hidden Markov Model (HMM) for de novo sequencing. The
main difference to standard methods which are all based on dynamic programming [2, 1]
lies in the fully probabilistic model. Our trained HMM defines a generative model for
mass spectra which, for instance, is used for scoring observed spectra according to their
likelihood given a peptide sequence. Besides predicting the most likely sequence, however,
the HMM framework is far more general in the sense that it additionally allows us to specify
the confidence in the predictions.
2
Tandem Mass Spectrometry
In a typical sequencing experiment by mass spectrometry a protein is digested with the
help of an enzyme. This digestion reaction breaks the protein into several peptides, each
of which consists of a short sequence of typically 10 to 20 amino acid residues, with an
additional H-atom at the N-terminus and an OH-group at the C-terminus.
DSRSCK
AYSARDGFSHEK
DGDGYASDZROPGFSHEK
1. MS
m/z
AYSARDGFSHEK
AYSAR
A
K
DGFSHEK
AY
EK
HEK
AYS
2. MS
m/z
Figure 1: In the first mass measurement the parent mass is selected. In the second measurement the peptide is dissociated and the mass of the ion fragments is measured.
There are two measurement steps in a tandem mass spectrometer. The first step is responsible for filtering peptides of a certain total mass (also called the parent mass). The difficulty
in measuring the parent mass arises from different 12 C/13 C isotope proportions of the
approximately 30-80 C-atoms contained in a peptide. Fluctuations of the 13 C fraction result in a binomial distribution of parent masses in the measurement. Given such an ?ion
count distribution? one can roughly estimate the mono-isotopic parent mass of the peptide,
where the term mono-isotopic here refers to a peptide that contains exclusively 12 C atoms.
In practice, all isotope configurations of a peptide with parent masses that do not exceed
the estimated mono-isotopic mass by more than a predefined offset are separated from the
other peptides and passed to the second spectrometer.
Figure 2: Top: The ideal peaks of a peptide sequence are drawn. Bottom: The spectrum of
the corresponding peptide.
In the second mass measurement, a peptide is split into two fragments by means of collision
induced dissociation with a noble gas. In almost all cases the peptide is broken between
two amino acids. Thus, an ideal spectrum is composed of the masses of all prefix and suffix
sequences of the peptide. Deviations from this ideal case are e.g. caused by problems in
determining the exact mono-isotopic mass of the fragments due to isotope shifts. Further
complications are caused by an accidental loss of water (H2 O), ammonia (NH3 ) or other
molecules in the collision step. Moreover, the ion counts are not uniformly distributed over
the spectrum. And last but not least, the measurements are noisy.
3
The Hidden Markov Model for de Novo Peptide Sequencing
A peptide can formally be described as a sequence of symbols from a fixed alphabet A of
20 amino acids. We will denote amino acids with ? ? A and the mass of an amino acid
with M (?). The input data is a spectrum of ion counts over all mass units. The ion count
for mass m is denoted by x(m). The spectra are discretized to approximately one Dalton
mass units and normalized such that the mean ion count per Dalton is constant.
The mono-isotopic parent mass m0p of the peptide P = (?1 , . . . , ?n ) with ?i ? A is
the sum of all amino acid masses plus a constant mass for the N- and C-termini. m 0p =
Pn
constN + i=1 M (?i ) + constC. For the sake of simplicity it is assumed that the N- and
C-termini are not present and thus the parent mass considered in the sequel is
Pn
(1)
mp = i=1 M (?i ) .
In the HMM framework a spectrum is regarded as a realization of a random process. The
physical process that generates spectra is based on the fact that a peptide is randomly broken
into two parts by interaction with a noble gas. Each of these parts is detected in the mass
spectrometer and increases the ion-count in the corresponding mass interval. Finally, a
histogram over many such events is measured. In order to derive a model of the generation
process, we make the simplifying assumptions that (i) breaks occur only at amino acid
boundaries, and (ii) the probability of observing a break after a certain amino acid depends
only on the amino acid itself. These assumptions allow us to model the generative process
by way of a Markov process on a finite state automation. In such a model, the process of
generating a spectrum for a peptide of known parent mass is formalized as a path through
the automaton in 1 Dalton steps until the constraint on the parent mass is satisfied.
3.1
Finite State Automaton
s0
sA
1
sA
2
sA
3
sR
1
sR
2
sR
3
sV
1
sV
2
sV
3
sA
71
s?
sR
156
s+
sV
99
Figure 3: The finite state machine of the Hidden Markov Model. For each amino acid ?
there is a list of M (?) states.
The finite state automaton (fig. 3) has one initial state s0 . For each amino acid ? ? A
?
there exists a list of M (?) states s?
1 , . . . , sM (?) . Together with the end states s+ and s?
the complete set of states is
S = {s0 } ? s?
(2)
j | ? ? A, 1 ? j ? M (?) ? {s+ , s? } .
The bold edges in the graph correspond to state transition probabilities a(s, t) from state s
to state t. Once the automation is in the first state s?
1 of a state list of one amino acid ?, it
has to pass through all other states within the specific list. Thus for the next M (?) steps
the list corresponding to amino acid ? is linearly traversed. If the automaton is in the last
?0
0
state s?
M (?) of a list, it can reach the start states s1 of any other amino acid ? . The random
variable for the state sequence is denoted by Y1 , . . . , Ymp . The transition probabilities are
a(s, t) = P {Yi+1
?
? 1
r?
= t|Yi = s} =
?
0
?
? ? ? A, 1 ? i < M (?) : s = s?
i ? t = si+1
?
?
? ? ? A, ? ? A : s = sm(?) ? t = s1
else .
(3)
The first row (a(s, t) = 1) describes the case where the automaton is in a non-terminating
state of a list of amino acid ? (1 ? i < M (?) : s = s?
i ), where the following state is
accepted with probability 1. The second row, on the contrary, refers to a terminating state
of a list. In such a case, the starting state of any other amino acid is selected with probability
r? . The probabilities r? are the probabilities of occurrence of amino acid ?.
The transition probabilities a(s0 , t) from the start state s0 are the occurrence probabilities
of the amino acids.
r? ? ? ? A : t = s ?
1
(4)
a(s0 , t) =
0
else
Finally one has to ensure that the parent mass constraint is fulfilled. In order to satisfy
the constraint we device a time dependent hidden Markov model in which the transition
probability changes with a heavy side function at time mp from a(s, t) to a0 (s, t). The
dotted arrows in figure 3 show the transition probabilities a0 (s, t) into the end states s+ and
s? .
?
? 1 ? ? ? A : s = s?
M (?) , t = s+
0
a (s, t) =
(5)
1 ? ? ? A, 1 ? i < M (?) : s = s?
i , t = s?
?
0 else
If the automaton is in the last state s?
M (?) of an amino acid state list, it changes to the
positive end state s+ with probability 1 since the parent mass constraint is satisfied. If the
automaton is in one of the other states, it changes to the negative end state s? since the
parent mass constraint is violated. It is important to realize that all amino acid sequences
that fulfill the parent mass constraint can be transformed into state sequences that end in
the positive state s+ and vice versa.
3.2
Emission Probabilities
0.1
0.1
0.35
0.09
0.09
0.3
0.08
0.08
0.07
0.25
0.07
0.06
0.06
0.2
0.05
H2O
0.05
NH 3
0.15
0.04
0.04
0.03
CO+NH 3
0.02
H2O+NH 3
0.03
0.1
CO
?40
NH 3
0.05
0.01
0.01
0
?50
H2O
0.02
2H2O
?30
?20
(a)
?10
0
10
0
0
?50
?50
?40
?40
?30
?30
?20
?20
?10
?10
0
0
10
10
(b)
Figure 4: Mean height of ion counts for different shifts with respect to the ideal prefix
fragments (a) and suffix fragments (b).
At each state of the finite state automaton an ion count value is emitted. Figure 4 shows
the mean ion count for different positions relative to the amino acid bound averaged over
all amino acids. The histograms are taken over the training examples described in the
experimental section. It happens quite frequently that an amino acid looses water (H 2 O)
or ammonia (NH3 ). The ion count patterns for the prefix fragments (fig. 4 a) and the
suffix fragments (fig. 4 b) are quite different due to chemical reasons. For instance, carbon
monoxide loss in the suffix fragments is an unlikely event. Suffix fragments are more stable
than prefix fragments: the central peak at position 0 (amino acid boundary) is three times
higher for the suffix fragments than for the prefix fragments. Note that in figure 4 b) we
used two different scales.
s+
m=m p
m=0
s0
Figure 5: Folding the spectrum in the middle makes the intern mirror symmetry of the
problem visible. The Markov chain models a sequence with three amino acids. The filled
circles correspond to the amino acid boundaries. Each amino acid bound generates an ion
count pattern for the prefix fragment and one for the suffix fragment.
Breaking a peptide in the second mass spectrometer produces both a prefix and a suffix
fragment. To simultaneously process peaks of both types of fragments, we use one forward
and one backward Markov chain which are independent of each other. Due to the inherent
mirror symmetry of the problem (fig. 5) it is sufficient to limit the length of both models to
mp /2. For the recognition process we assume that we simultaneously observe two peaks
xm,1 = x(m) and xm,2 = x(mp ? m) in step m. The joint observation of the prefix and
the suffix peaks is an essential modeling step in our method.
The forward and the backward Markov chains are extended to hidden Markov models to
describe the ion counts in the mass spectra. The emission probabilities depend on the two
states of the prefix and suffix sequence, since these states give rise to ion counts in the
measurements. We define
? m = xm = (x(m), x(mp ? m)) | Y?m = (s, s0 )
bs,s0 (xm ) = P X
(6)
as the emission probabilities of ion counts.
? m are the (coupled) random variables of the ion counts. The hidden variables for the state
X
? m describes the transition
sequence are denoted by Y?m . This notion of coupled variables X
from two independent Markov chains to one coupled hidden Markov model with a squared
number of states (2-tuple states).
The joint probability of observable and hidden variables given the parent mass m p is
P {X = x, Y = y | s+ , mp } = a (s0 , y1 ) a0 ymp , s+ ?
(7)
? m ?1
?
p
2
?
? Y
??
bym ,ymp ?m (xm )a (ym , ym+1 ) a ymp ?m , ymp ?m+1 ? a y mp ?1 , y mp ?1 +1
m=1
2
2
This formula holds for parent masses with an odd Dalton value, an equivalent formula can
be derived for the even case. The first term in eq. (7) is the joint probability from s 0 to
y1 in the prefix model and the transition ymp to s+ in the suffix model. In each term of
the product, two peaks are observed on both sides of the spectrum: one at position m and
the other at the mirror position mp ? m. The joint probability of emissions is defined by
bym ,ymp ?m (xm , xmp ?m ). Furthermore, the transition probabilities of the prefix and suffix
sequences are multiplied which reflects the independence assumption of the Markov model.
The two chains are connected by the transition probability a(y(mp ?1)/2 , y(mp ?1)/2+1 ) of
traversing from the last state of the forward Markov chain to the first state of the backward
chain.
3.3
Most Probable Sequence
The input spectrum usually comes with an estimate of the parent mass with a tolerance of
about 1 Dalton. Using a maximum likelihood approach the parent mass estimate is
X
P {X = x, Y = y | s+ , mp } . (8)
m
? p = argmax P {X = x | s+ , mp } = argmax
mp
mp
y
The sum over all sequences can be computed efficiently by dynamic programming using
the forward algorithm.
One result of de novo peptide sequencing is the computation of the best sequence generating a given spectrum. Given the estimated parent mass m
? p the maximum posterior estimate
of the sequence is
y ? = argmax P {Y = y | X = x, s+ , m
? p } = argmax P {X = x, Y = y | s+ , m
? p} .
y
y
(9)
The best sequence can efficiently be found by the Viterbi algorithm. To compute the posterior probability one has to normalize the joint probability P {X = x, Y = y | s+ , m
? p } by
the evidence P {X = x | s+ , m
? p } using the forward-backward algorithm.
In the mass spectra ions with very low mass or almost parent mass are less frequently
observed than ions with a medium mass. Therefore it becomes quite difficult to estimate the
whole sequence with a high score. It is also possible to give a score for each subsequence
of the peptide, especially a score for each amino acid. An amino acid is a subsequence
yp , . . . , yq of the state sequence y1 , . . . , ymp .
=
P {yp , . . . , yq | s+ , x, mp }
P
P
y1 ,...yp?1
yq+1 ,...,ym P y1 , . . . , ymp , x | s+ , mp
p
P {x | s+ , mp }
(10)
(11)
This can again be computed by some variation of the forward and backward algorithm.
3.4
Simplification of the Model
The coupled hidden Markov model has 2 3752 = 5 640 625 states that leads to a runtime
of 20 minutes per peptide which for practical applications is problematic. A significant
simplification is achieved by assuming that there are two spectra observed, where the second one is the mirror version of the first one. The emission probabilities in this simplified
model only depend on the states of the prefix Markov chain (fig. 6). Thus the emission of
mirror peaks x(mp ?m) is deterministically coupled to the emission of the peak xm . Since
this model has only 2 375 states, the computation time reduces to 1-2 seconds per peptide.
4
Experiments
In our experiments a protein probe of plant cell vacuoles (Arabidopsis thaliana) was digested with trypsin. The mass spectrometer gave an output of 7056 different candidate spectra. From a database search with SEQUEST [3] and further validation with
PeptideProphet [4], 522 spectra with a confidence larger than 90% were extracted.
It was shown that the PeptideProphet score is a very reliable scoring method for peptide identification by database search. The database output was used as training data. The
m=m p
m=0
m=m p
m=0
s0
s+
Figure 6: In the simplified model two mirrored spectra are observed. The emission of
symbols is coupled with the amino acid bounds of the prefix sequence.
quality of the HMM inference is measured by the ratio of common amino acid boundaries
and the number of amino acids in the database sequence. The performance of the HMM
was tested by leave-one-out cross validation: in each training step the emission probabilities and the amino acid occurrence probabilities are re-estimated, with one sequence
excluded from the training set. To estimate the emission probabilities, the ion count is discretized to a fixed number of bins, in such a way that all bins contain an equal number of
counts. The leave-one-out scheme is repeated for different numbers of discretization levels.
1
0.9
0.8
0.7
recall
0.6
0.5
0.4
0.3
0.2
0.1
0
2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30
number of bins
Figure 7: Cross validation of recall rates for different number of bins in the discretization
process. Depicted are the lower quartile, the median and the upper quartile.
The resulting performance recall rate are depicted in figure 7. Choosing 5 bins yields the
highest recall value.
We have chosen the prominent de novo sequencing programs LUTEFISK [6] and PEAKS
[5] as competitors for the simplified HMM. We compared the sequence from the HMM
with the highest scoring sequences from the other programs. In figure 8 a) the estimated
parent masses compared to the database parent mass is drawn. The plot demonstrates that
all de novo sequencing methods tend to overestimate the parent mass. The best one is the
HMM with 89.1% correct estimations, whereas only 59.3% of the LUTEFISK estimates
and 58.1% of the PEAKS estimates are correct. In figure 8 b) boxplots of the recognition
rate of peak positions is drawn. The three lines in the box correspond to the lower quartile,
the median and the upper quartile of the distribution. The median recall of the HMM is
75.0%, for Lutefisk 53.9% and for Peaks 56.7%. Note that the lower quartile of the HMM
results is above 50%, whereas it is below 10% for the other programs.
5
Conclusion and Further Work
A novel method for the analysis of mass spectra in de novo peptide sequencing is presented
in this paper. The proposed hidden Markov model is a fully probabilistic model for the
generation process of mass spectra. The model was tested on mass spectra from vacuola
1
1
1
0.9
0.9
0.9
1
0.9
0.8
0.8
0.8
0.7
0.7
0.7
0.6
0.6
0.6
0.5
0.5
0.5
0.4
0.4
0.4
0.3
0.3
0.3
0.2
0.2
0.2
0.2
0.1
0.1
0.1
0.1
0.8
0.7
Recall
0.6
0.5
0.4
0
<?2 ?2 ?1 0 1 2 3
Lutefisk
0
?3 ?2 ?1 0 1 2 3
Peaks
(a)
0
0.3
0
?3 ?2 ?1 0 1 2 3
HMM
Lutefisk
Peaks
HMM
(b)
Figure 8: a) Histogram on difference of estimated parent mass and database output. b)
Recall of peak positions.
proteins. The HMM clearly outperforms its competitors in recognition of the parent mass
and peak localization. In further work additional model parameters will be introduced to
represent and to detect amino acids with post-translational modifications. Reliable subsequences can further be used for a tagged database search to identify peptides with posttranslational modifications. Our method shows a large potential for high throughput de
novo sequencing of proteins which is unmatched by competing techniques.
Acknowledgment This work has been partially supported by DFG grant # Buh 914/5.
References
[1] Sacha Baginsky, Mark Cieliebak, Wilhelm Gruissem, Torsten Kleffmann, Zsuzsanna
Lipt?ak, Matthias M?uller, and Paolo Penna. Audens: A tool for automatic de novo
peptide sequencing. Technical Report 383, ETH Zurich, Dept. of Computer Science,
2002.
[2] Ting Chen, Ming-Yang Kao, Matthew Tepel, John Rush, and George M. Church. A
dynamic programming approach to de novo peptide sequencing via tandem mass spectrometry. Journal of Computational Biology, 8(3):325?337, 2001.
[3] Jimmy K. Eng, Ashley L. McCormack, and John R. Yates. An approach to correlate
tandem mass spectral data of peptides with amino acid sequences in a protein database.
American Society for Mass Spectrometry, 5(11):976?989, 1994.
[4] Andrew Keller, Alexey I. Nesvizhskii, Eugene Kolker, and Ruedi Aebersold. Empirical
statistical model to estimate the accuracy of peptide identifications made by MS/MS
and database search. Analytical Chemistry, 2002.
[5] Bin Ma, Kaizhong Zhang, Christopher Hendrie, Chengzhi Liang, Ming Li, Amanda
Doherty-Kirby, and Gilles Lajoie. Peaks: Powerful software for peptide de novo sequencing by tandem mass spectrometry. Rapid Communication in Mass Spectrometry,
17(20):2337?2342, 2003.
[6] J. Alex Taylor and Richard S. Johnson. Implementation and uses of automated de novo
peptide sequencing by tandem mass spectrometry. Analytical Chemistry, 73:2594?
2604, 2001.
| 2668 |@word torsten:1 version:1 middle:1 proportion:1 open:1 simplifying:1 eng:1 initial:1 configuration:1 contains:1 fragment:16 exclusively:1 score:4 terminus:4 prefix:13 outperforms:2 reaction:1 current:2 discretization:2 surprising:1 si:1 john:2 realize:1 visible:1 plot:1 generative:2 selected:2 device:1 short:1 complication:1 zhang:1 height:1 jonas:1 consists:1 rapid:1 roughly:1 frequently:2 discretized:2 ming:2 ruedi:1 tandem:6 becomes:1 moreover:1 mass:64 medium:1 loos:1 runtime:1 demonstrates:1 arabidopsis:1 unit:2 grant:1 overestimate:1 positive:2 limit:1 ak:1 fluctuation:1 path:1 approximately:2 plus:1 alexey:1 challenging:2 co:2 limited:1 hek:1 averaged:1 practical:2 responsible:1 acknowledgment:1 practice:1 empirical:1 eth:4 significantly:1 matching:1 confidence:2 refers:2 protein:7 proteome:1 equivalent:1 roth:1 starting:1 jimmy:1 automaton:8 keller:1 simplicity:1 formalized:1 regarded:1 oh:1 notion:1 variation:1 exact:1 programming:3 us:1 hypothesis:1 recognition:3 ymp:9 database:10 observed:5 bottom:1 spectrometer:5 connected:1 highest:2 broken:2 dynamic:3 trained:1 terminating:2 depend:2 localization:1 joint:5 alphabet:1 separated:1 describe:1 detected:1 choosing:1 quite:3 larger:1 reconstruct:2 novo:18 fischer:2 noisy:1 itself:1 sequence:30 matthias:1 analytical:2 interaction:1 product:1 realization:1 aebersold:1 kao:1 normalize:1 ammonia:2 parent:25 produce:1 generating:2 leave:2 help:1 derive:1 andrew:1 measured:3 odd:1 eq:1 sa:4 come:1 switzerland:3 correct:2 quartile:5 bin:6 probable:1 traversed:1 hold:1 considered:1 ays:1 viterbi:1 matthew:1 estimation:1 peptide:37 vice:1 dalton:5 tool:1 reflects:1 uller:1 clearly:1 fulfill:1 pn:2 volker:1 derived:1 emission:10 joachim:1 sequencing:20 likelihood:2 suppression:1 detect:2 sense:1 inst:1 inference:1 dependent:1 suffix:12 typically:1 unlikely:1 a0:3 hidden:12 transformed:1 ashley:1 translational:1 denoted:3 equal:1 once:1 atom:3 biology:1 throughput:1 noble:2 report:1 inherent:1 richard:1 randomly:1 composed:1 simultaneously:2 dfg:1 argmax:4 chain:8 predefined:1 edge:1 tuple:1 traversing:1 filled:1 taylor:1 circle:1 re:1 rush:1 theoretical:1 instance:2 modeling:1 measuring:1 deviation:1 sacha:2 johnson:1 sv:4 peak:19 sequel:1 probabilistic:2 together:1 ym:3 xmp:1 squared:1 central:1 satisfied:2 again:1 unmatched:1 ek:1 american:1 yp:3 li:1 potential:1 de:18 chemistry:2 bold:1 automation:2 satisfy:1 caused:2 mp:19 depends:1 break:3 observing:1 start:2 accuracy:1 acid:36 efficiently:2 dissociation:1 correspond:3 yield:1 identify:1 identification:2 penna:1 reach:1 competitor:2 recall:7 higher:1 specify:1 box:1 furthermore:1 until:1 christopher:1 lack:1 defines:1 quality:2 normalized:1 contain:1 tagged:1 chemical:1 excluded:1 isotope:3 m:4 prominent:1 ay:1 complete:1 demonstrate:1 doherty:1 isotopic:5 novel:2 common:1 physical:1 nh:4 measurement:8 significant:1 versa:1 automatic:1 digested:2 stable:1 enzyme:1 posterior:2 inf:1 certain:2 yi:2 scoring:3 additional:2 george:1 spectrogram:1 ii:1 reduces:1 technical:1 cross:2 post:1 prediction:1 histogram:3 represent:1 achieved:1 cell:1 ion:19 folding:1 spectrometry:9 residue:1 whereas:2 interval:1 else:3 median:3 sr:4 induced:1 tend:1 contrary:1 emitted:1 yang:1 ideal:4 exceed:1 split:1 automated:1 independence:1 gave:1 competing:1 shift:2 passed:1 suffer:1 peter:1 collision:2 dna:1 exist:1 mirrored:1 problematic:1 dotted:1 estimated:5 fulfilled:1 per:3 yates:1 paolo:1 group:1 drawn:3 mono:5 boxplots:1 backward:5 graph:1 fraction:1 sum:2 powerful:1 almost:2 thaliana:1 bound:3 simplification:2 accidental:1 occur:1 constraint:6 alex:1 software:1 sake:1 generates:2 buh:1 trypsin:1 according:1 describes:2 kirby:1 b:1 s1:2 happens:1 modification:2 restricted:1 taken:1 zurich:7 count:17 end:5 available:1 highthroughput:1 multiplied:1 probe:1 observe:1 spectral:1 occurrence:3 bym:2 binomial:1 top:1 ensure:1 ting:1 especially:1 society:1 hmm:14 lajoie:1 water:2 reason:1 assuming:1 besides:1 length:1 wilhelm:2 ratio:1 liang:1 difficult:1 carbon:1 negative:1 rise:1 implementation:1 gilles:1 upper:2 observation:1 markov:18 sm:2 finite:5 gas:2 extended:1 communication:1 y1:6 introduced:1 bernd:2 amanda:1 usually:2 pattern:2 xm:7 below:1 program:3 reliable:3 event:2 difficulty:1 predicting:1 buhmann:1 scheme:1 yq:3 church:1 coupled:6 eugene:1 determining:1 relative:1 plant:2 fully:2 grossmann:1 loss:2 generation:2 filtering:1 validation:3 h2:1 sufficient:1 s0:11 heavy:1 row:2 supported:1 last:4 side:2 allow:1 institute:2 distributed:1 tolerance:1 overcome:1 boundary:4 transition:9 genome:1 forward:6 made:1 franz:1 simplified:4 far:1 correlate:1 observable:1 roos:1 vacuole:1 assumed:1 spectrum:25 subsequence:3 search:4 additionally:1 molecule:1 symmetry:2 dissociated:1 main:1 linearly:1 arrow:1 whole:1 repeated:1 amino:36 fig:5 precision:1 position:6 deterministically:1 lie:1 candidate:1 breaking:1 formula:2 minute:1 specific:1 symbol:2 offset:1 list:9 evidence:1 exists:1 essential:1 effectively:1 mirror:5 lipt:1 chen:1 depicted:2 likely:1 intern:1 contained:2 partially:1 ch:4 extracted:1 ma:1 goal:1 change:3 typical:1 uniformly:1 total:1 specie:1 called:1 pas:1 accepted:1 experimental:1 formally:1 mark:1 arises:1 ethz:1 violated:1 dept:1 tested:2 |
1,837 | 2,669 | Maximal Margin Labeling for Multi-Topic Text
Categorization
Hideto Kazawa, Tomonori Izumitani, Hirotoshi Taira and Eisaku Maeda
NTT Communication Science Laboratories
Nippon Telegraph and Telephone Corporation
2-4 Hikaridai, Seikacho, Sorakugun, Kyoto 619-0237 Japan
{kazawa,izumi,taira,maeda}@cslab.kecl.ntt.co.jp
Abstract
In this paper, we address the problem of statistical learning for multitopic text categorization (MTC), whose goal is to choose all relevant topics (a label) from a given set of topics. The proposed algorithm, Maximal Margin Labeling (MML), treats all possible labels as independent
classes and learns a multi-class classifier on the induced multi-class categorization problem. To cope with the data sparseness caused by the huge
number of possible labels, MML combines some prior knowledge about
label prototypes and a maximal margin criterion in a novel way. Experiments with multi-topic Web pages show that MML outperforms existing
learning algorithms including Support Vector Machines.
1
Multi-topic Text Categorization (MTC)
This paper addresses the problem of learning for multi-topic text categorization (MTC),
whose goal is to select all topics relevant to a text from a given set of topics. In MTC,
multiple topics may be relevant to a single text. We thus call a set of topics label, and say
that a text is assigned a label, not a topic.
In almost all previous text categorization studies (e.g. [1, 2]), the label is predicted by
judging each topic?s relevance to the text. In this decomposition approach, the features
specific to a topic, not a label, are regarded as important features. However, the approach
may result in inefficient learning as we will explain in the following example.
Imagine an MTC problem of scientific papers where quantum computing papers are assigned multi-topic label ?quantum physics (QP) & computer science (CS)?. (QP and CS
are topics in this example.) Since there are some words specific to quantum computing
such as ?qbit1 ?, one can say that efficient MTC learners should use such words to assign
label QP & CS. However, the decomposition approach is likely to ignore these words since
they are only specific to a small portion of the whole QP or CS papers (there are many more
QP and CS papers than quantum computing papers), and therefore are not discriminative
features for either topic QP or CS.
1
Qbit is a unit of quantum information, and frequently appears in real quantum computing literatures, but rarely seen in other literatures.
Symbol
x(? Rd )
t1 , t2 , . . . , tl
T
L, ?(? T )
L[j]
?(= 2T )
{(xi , Li )}m
i=1
Meaning
A document vector
Topics
The set of all topics
A label
The binary representation of L. 1 if tj ? L and 0 otherwise.
The set of all possible labels
Training samples
Table 1: Notation
Parametric Mixture Model (PMM) [3] adopts another approach to MTC. It is assumed in
PMM that multi-topic texts are generated from a mixture of topic-specific word distributions. Its decision on labeling is done at once, not separately for each topic. However, PMM
also has a problem with multi-topic specific features such as ?qbit? since it is impossible
for texts to have such features given PMM?s mixture process.
These problems with multi-topic specific features are caused by dependency assumptions
between labels, which are explicitly or implicitly made in existing methods. To solve these
problems, we propose Maximal Margin Labeling, which treats labels as independent
classes and learns a multi-class classifier on the induced multi-class problem.
In this paper, we first discuss why multi-class classifiers cannot be directly applied to MTC
in Section 2. We then propose MML in Section 3, and address implementation issues in
Section 4. In Section 5, MML is experimentally compared with existing methods using a
collection of multi-topic Web pages. We summarize this paper in Section 6.
2
Solving MTC as a Multi-Class Categorization
To discuss why existing multi-class classifiers do not work in MTC, we start from the
multi-class classifier proposed in [4]. Hereafter we use the notation given in Table 1. The
multi-class classifier in [4] categorizes an object into the class whose prototype vector is
the closest to the object?s feature vector. By substituting label for class, the classifier can
be written as follows.
f (x) = arg max hx, m? iX
(1)
???
where h, iX is the the inner product of Rd , and m? ? Rd is the prototype vector of label ?.
Following the similar argument as in [4], the prototype vectors are learned by solving the
following maximal margin problem2 .
X
X
1
min kM k2 + C
?i?
M 2
1?i?m ???,?6=Li
s.t.
hxi , mLi iX ? hxi , m? iX ? 1 ? ?i?
for 1 ? i ? m, ?? 6= Li ,
(2)
where M is the prototype matrix whose columns are the prototype vectors, and kM k is the
Frobenius matrix norm of M .
Note that Eq. (1) and Eq. (2) cover only training samples? labels, but also all possible labels.
This is because the labels unseen in training samples may be relevant to test samples. In
2
In Eq.(2), we penalize all violation of the margin constraints. On the other hand, Crammer
and Singer penalize only the largest violation of the margin constraint for each training sample [4].
We chose the ?penalize-all? approach since it leads to an optimization problem without equality
constraints (see Eq.(7)), which is much easier to solve than the one in [4].
usual multi-class problems, such unseen labels seldom exist. In MTC, however, the number
of labels is generally very large (e.g. one of our datasets has 1,054 labels (Table 2)), and
unseen labels often exist. Thus it is necessary to consider all possible labels in Eq. (1) and
Eq. (2) since it is impossible to know which unseen labels are present in the test samples.
There are two problems with Eq. (1) and Eq. (2). The first problem is that they involve the
prototype vectors of seldom or never seen labels. Without the help of prior knowledge about
where the prototype vectors should be, it is impossible to obtain appropriate prototype
vectors for such labels. The second problem is that these equations are computationally too
demanding since they involve combinatorial maximization and summation over all possible
labels, whose number can be quite large. (For example, the number is around 230 in the
datasets used in our experiments.) We will address the first problem in Section 3 and the
second problem in Section 4.
3
Maximal Margin Labeling
In this section, we incorporate some prior knowledge about the location of prototype vectors into Eq. (1) and Eq. (2), and propose a novel MTC learning algorithm, Maximal
Margin Labeling (MML).
As prior knowledge, we simply assume that the prototype vectors of similar labels should
be placed close to each other. Based on this assumption, we first rewrite Eq. (1) to yield
f (x) = arg max hM T x, e? iL ,
(3)
???
where h, iL is the inner product of R|?| and {e? }??? is the orthonormal basis of R|?| . The
classifier of Eq. (3) can be interpreted as a two-step process: the first step is to map the
vector x into R|?| by M T , and the second step is to find the closest e? to image M T x.
Then we replace {e? }??? with (generally) non-orthogonal vectors {?(?)}??? whose geometrical configuration reflects label similarity. More formally speaking, we use vectors
{?(?)}??? that satisfy the condition
h?(?1 ), ?(?2 )iS = S(?1 , ?2 ) for ??1 , ?2 ? ?,
(4)
where h, iS is an inner product of the vector space spanned by {?(?)}??? , and S is a
Mercer kernel [5] on ? ? ? and is a similarity measure between labels. We call the vector
space spanned by {?(?)} VS .
With this replacement, MML?s classifier is written as follows.
f (x) = arg max hW x, ?(?)iS ,
(5)
???
where W is a linear map from Rd to VS . W is the solution of the following problem.
m
X
X
1
min
kW k2 + C
?i?
W
2
i=1 ???,?6=Li
?(Li )??(?)
s.t.
W xi ,
? 1??i? , ?i? ? 0 for 1 ? i ? m, ?? 6= Li . (6)
k?(Li )??(?)k
Note that if ?(?) is replaced by e? , Eq. (6) becomes identical to Eq. (2) except for a scale
factor. Thus Eq. (5) and Eq. (6) are natural extensions of the multi-class classifier in [4].
We call the MTC classifier of Eq. (5) and Eq. (6) ?Maximal Margin Labeling (MML)?.
Figure 1 explains the margin (the inner product in Eq. (6)) in MML. The margin represents
the distance from the image of the training sample xi to the boundary between the correct
label Li and wrong label ?. MML optimizes the linear map W so that the smallest margin
between all training samples and all possible labels becomes maximal, along with a penalty
C for the case that samples penetrate into the margin.
Figure 1: Maximal Margin Labeling
Dual Form For numerical computation, the following Wolfe dual form of Eq. (6) is more
convenient. (We omit its derivation due to space limits.)
X
S(Li , Li? )?S(Li , ?? )?S(?, Li? )+S(?, ?? )
1 X X ? ??
p
?i ?i? (xi ?xi? )
max
?i? ?
2
??
2 (1?S(Li , ?))(1?S(Li? , ?? ))
i
?
?
i,? i ,?
i,?
s.t. 0 ? ?i? ? C for 1 ? i ? m, ?? 6= Li ,
(7)
Pm P
P
?
where we denote i=1 ???,??6=Li by i,? , and ?i are the dual variables corresponding
to the first inequality constraints in Eq. (6). Note that Eq. (7) does not contain ?(?): all the
computations involving ? can be done through the label similarity S. Additionally xi only
appears in the inner products, and therefore can be replaced by any kernel of x.
Using the solution ?i? of Eq. (7), the MML?s classifier in Eq. (5) can be written as follows.
X
S(Li , L)?S(?, L)
.
(8)
f (x) = arg max
?i? (x?xi ) p
2(1?S(Li , ?))
L??
i,?
Label Similarity3 As examples of label similarity, we use two similarity measures: Dice
measure and cosine measure.
4
Dice measure
Cosine measure
4
4.1
Pl
2 j=1 ?1 [j]?2 [j]
2|?1 ??2 |
SD (?1 , ?2 ) =
= Pl
.
(9)
Pl
|?1 |+|?2 |
j=1 ?1 [j] +
j=1 ?2 [j]
Pl
|?1 ? ?2 |
j=1 ?1 [j]?2 [j]
p
qP
= qP
SC (?1 , ?2 ) = p
.(10)
l
l
|?1 | |?2 |
j=1 ?1 [j]
j=1 ?2 [j]
Efficient Implementation
Approximation in Learning
Eq. (7) contains the sum over all possible labels. As the number of topics (l) increases, this
summation rapidly becomes intractable since |?| grows exponentially as 2l . To circumvent
3
The following discussion is easily extended to include the case that both ?1 and ?2 are empty
although we do not discuss the case due to space limits.
this problem, we approximate the sum over all possible labels in Eq. (7) by the partial sum
over ?i? of |(A ? B c ) ? (Ac ? B)| = 1 and set all the other ?i? to zero. This approximation
reduces the burden of the summation quite a lot: the number of summands is reduced from
2l to l, which is a huge reduction especially when many topics exist.
To understand the rationale behind the approximation, first note that ?i? is the dual variable
corresponding to the first inequality constraint (the margin constraint) in Eq. (7). Thus ?i?
is non-zero if and only if W xi falls in the margin between ?(Li ) and ?(?). We assume
that this margin violation mainly occurs when ?(?) is ?close? to ?(Li ), i.e. |(A ? B c ) ?
(Ac ? B)| = 1. If this assumption holds well, the proposed approximation of the sum will
lead to a good approximation of the exact solution.
4.2
Polynomial Time Algorithms for Classification
The classification of MML (Eq. (8)) involves the combinatorial maximization over all possible labels, so it can be a computationally demanding process. However, efficient classification algorithms are available when either the cosine measure or dice measure is used as
label similarity.
Eq. (8) can be divided into the subproblems by the number of topics in a label.
f (x) =
arg max g(x, L),
?n
L
(11)
? 1 ,L
? 2 ,...,L
?l}
L?{L
=
arg max g(x, L).
(12)
L??,|L|=n
where g(x) is
g(x, L)
=
l
X
cn [j]L[j],
j=1
cn [j]
=
? P
??
2Li [j]
2?[j]
i (x?xi )
?
?
.
?
?
i,?
|?|+n
2(1?SD (Li ,?)) |Li |+n
?
P
?
(x?x
)
?[j]
L
[j]
i
i
?
i
?
?
?
?
?
? ?
i,?
2(1?SC (Li ,?))
|Li | n
|?| n
if SD is used.
(13)
if SC is used.
Here n = |L|. The computational cost of Eq. (13) for all j is O(n? l) (n? is the number of
non-zero ?), and that of Eq. (12) is O(l log l). Thus the total cost of the classification by
Eq. (11) is O(n? l2 + l2 log l). On the other hand, n? is O(ml) under the approximation
described above. Therefore, the classification can be done within O(ml3 ) computational
steps, which is a significant reduction from the case that the brute force search is used in
Eq. (8).
5
Experiments
In this section, we report experiments that compared MML to PMM [3], SVM5 [6], and
BoosTexter [2] using a collection of Web pages. We used a normalized linear kernel
k(x, x? ) = x ? x? /kxkkx? k in MML and SVM. As for BoosTexter, ?real abstaining AdaBoost.MH? was used as the weak learner.
5.1
Experimental Setup
The datasets used in our experiment represent the Web page collection used in [3] (Table 2). The Web pages were collected through the hyperlinks from Yahoo!?s top directory
5
For each topic, an SVM classifier is trained to predict whether the topic is relevant (positive) or
irrelevant (negative) to input doucments.
Dataset Name (Abbrev.)
Arts & Humanities (Ar)
Business & Economy (Bu)
Computers & Internet (Co)
Education (Ed)
Entertainment (En)
Health (He)
Recreation (Rc)
Reference (Rf)
Science (Si)
Social Science (SS)
Society & Culture (SC)
#Text
#Voc
7,484
11,214
12,444
12,030
12,730
9,205
12,828
8,027
6,428
12,111
14,512
23,146
21,924
34,096
27,534
32,001
30,605
30,324
39,679
37,187
52,350
31,802
#Tpc #Lbl
26
30
33
33
21
32
22
33
40
39
27
Label Size Frequency (%)
1
2
3
4 ?5
599 55.6 30.5 9.7 2.8 1.4
233 57.6 28.8 11.0 1.7 0.8
428 69.8 18.2 7.8 3.0 1.1
511 66.9 23.4 7.3 1.9 0.6
337 72.3 21.1 4.5 1.0 1.1
335 53.2 34.0 9.5 2.4 0.9
530 69.2 23.1 5.6 1.4 0.6
275 85.5 12.6 1.5 0.3 0.1
457 68.0 22.3 7.3 1.9 0.5
361 78.4 17.0 3.7 0.7 0.3
1054 59.6 26.1 9.2 2.9 2.2
Table 2: A summary of the web page datasets. ?#Text? is the number of texts in the dataset,
?#Voc? the number of vocabularies (i.e. features), ?#Tpc? the number of topics, ?#Lbl? the
number of labels, and ?Label Size Frequency? is the relative frequency of each label size.
(Label size is the number of topics in a label.)
Method
MML
PMM
SVM
Boost
Feature Type
TF, TF?IDF
TF
TF, TF?IDF
Binary
Parameter
C = 0.1, 1, 10
Model1, Model2
C = 0.1, 1, 10
R = {2, 4, 6, 8, 10}?103
Table 3: Candidate feature types and learning parameters. (R is the number of weak hypotheses.) The underlined fetures and parameters were selected for the evaluation with the
test data.
(www.yahoo.com), and then divided into 11 datasets by Yahoo?s top category. Each
page is labeled with the Yahoo?s second level sub-categories from which the page is hyperlinked. (Thus, the sub-categories are topics in our term.) See [3] for more details about
the collection. Then the Web pages were converted into three types of feature vectors: (a)
Binary vectors, where each feature indicates the presence (absence) of a term by 1 (0); (b)
TF vectors, where each feature is the number of appearances of a term (term frequency);
and (c) TF?IDF vectors, where each feature is the product of term frequency and inverse
document frequency [7].
To select the best combinations of feature types and learning parameters such as the penalty
C for MML, the learners were trained on 2,000 Web pages with all combinations of feature and parameter listed in Table 3, and then were evaluated by labeling F-measure on
independently drawn development data. The combinations which achieve the best labeling
F-measures (underlined in Table 3) were used in the following experiments.
5.2
Evaluation Measures
We used three measures to evaluate labeling performance: labeling F-measure, exact match
ratio, and retrieval F-measure. In the following definitions, {Lpred
}ni=1 and {Ltrue
}ni=1
i
i
mean the predicted labels and the true labels, respectively.
Labeling F-measure Labeling F-measure FL evaluates the average labeling performance
while taking partial match into account.
Pl
pred
n
n
true
? Ltrue
|
1 X 2 j=1 Li [j]Li [j]
1 X 2|Lpred
i
i
FL =
=
.
(14)
P
n i=1 |Lpred
n i=1 lj=1 (Lpred
| + |Ltrue
|
[j] + Ltrue
[j])
i
i
i
i
Dataset
Ar
Bu
Co
Ed
En
He
Rc
Rf
Si
SS
SC
Avg
MD
0.55
0.80
0.62
0.56
0.64
0.74
0.63
0.67
0.61
0.73
0.60
0.65
Labeling F-measure
MC PM SV BO
0.44 0.50 0.46 0.38
0.81 0.75 0.76 0.75
0.59 0.61 0.55 0.47
0.43 0.51 0.48 0.37
0.52 0.61 0.54 0.49
0.74 0.66 0.67 0.60
0.46 0.55 0.49 0.44
0.58 0.63 0.56 0.50
0.54 0.52 0.47 0.39
0.71 0.66 0.64 0.59
0.55 0.54 0.49 0.44
0.58 0.59 0.56 0.49
MD
0.44
0.63
0.51
0.45
0.55
0.58
0.54
0.60
0.52
0.65
0.44
0.54
Exact Match Ratio
MC PM SV
0.32 0.21 0.29
0.62 0.48 0.57
0.46 0.35 0.41
0.34 0.19 0.30
0.44 0.31 0.42
0.53 0.34 0.47
0.38 0.25 0.37
0.51 0.39 0.49
0.43 0.22 0.36
0.60 0.45 0.55
0.40 0.21 0.32
0.46 0.31 0.41
BO
0.22
0.53
0.34
0.23
0.36
0.39
0.33
0.41
0.28
0.49
0.27
0.35
MD
0.30
0.25
0.27
0.25
0.37
0.35
0.47
0.29
0.37
0.36
0.29
0.32
Retrieval F-measure
MC PM SV BO
0.26 0.24 0.29 0.22
0.27 0.20 0.29 0.20
0.25 0.19 0.30 0.17
0.23 0.21 0.25 0.16
0.33 0.30 0.35 0.29
0.35 0.23 0.35 0.26
0.39 0.36 0.40 0.33
0.25 0.24 0.25 0.16
0.35 0.28 0.31 0.19
0.35 0.18 0.31 0.15
0.28 0.25 0.26 0.20
0.30 0.24 0.31 0.21
Table 4: The performance comparison by labeling F-measure (left), exact match ratio (middle) and retrieval F-measure (right). The bold figures are the best ones among the five
methods, and the underlined figures the second best ones. MD, MC, PM, SV, and BO
represent MML with SD , MML with SC , PMM, SVM and BoosTexter, respectively.
Exact Match Ratio Exact match ratio EX counts only exact matches between the predicted label and the true label.
n
EX =
1X
I[Lpred
= Ltrue
],
i
i
n i=1
(15)
where I[S] is 1 if the statement S is true and 0 otherwise.
Retrieval F-measure6 For real tasks, it is also important to evaluate retrieval performance, i.e. how accurately classifiers can find relevant texts for a given topic. Retrieval
F-measure FR measures the average retrieval performance over all topics.
Pn
l
[j]Ltrue
[j]
1 X 2 i=1 Lpred
i
i
FR =
.
(16)
Pn
true
l j=1 i=1 (Lpred
[j] + Li [j])
i
5.3
Results
First we trained the classifiers with randomly chosen 2,000 samples. We then calculated
the three evaluation measures for 3,000 other randomly chosen samples. This process was
repeated five times, and the resulting averaged values are shown in Table 4. Table 4 shows
that the MMLs with Dice measure outperform other methods in labeling F-measure and
exact match ratio. The MMLs also show the best performance with regard to retrieval Fmeasure although the margins to the other methods are not as large as observed in labeling
F-measure and exact match ratio. Note that no classifier except MML with Dice measure
achieves good labeling on all the three measures. For example, PMM shows high labeling
F-measures, but its performance is rather poor when evaluated in retrieval F-measure.
As the second experiment, we evaluated the classifiers trained with 250?2000 training samples on the same test samples. Figure 2 shows each measure averaged over all datasets. It is
observed that the MMLs show high generalization even when training data is small. An interesting point is that MML with cosine measure achieves rather high labeling F-measures
and retrieval F-measure with training data of smaller size. Such high-performace, however,
does not continue when trained on larger data.
6
FR is called ?the macro average of F-measures? in the text categorization community.
Figure 2: The learning curve of labeling F-measure (left), exact match ratio (middle) and
retrieval F-measure (right). MD, MC, PM, SV, BO mean the same as in Table 4.
6
Conclusion
In this paper, we proposed a novel learning algorithm for multi-topic text categorization.
The algorithm, Maximal Margin Labeling, embeds labels (sets of topics) into a similarityinduced vector space, and learns a large margin classifier in the space. To overcome the
demanding computational cost of MML, we provide an approximation method in learning
and efficient classification algorithms. In experiments on a collection of Web pages, MML
outperformed other methods including SVM and showed better generalization.
Acknowledgement
The authors would like to thank Naonori Ueda, Kazumi Saito and Yuji Kaneda of Nippon
Telegraph and Telephone Corporation for providing PMM?s codes and the datasets.
References
[1] Thorsten Joachims. Text categorization with support vector machines: learning with
many relevant features. In Claire N?edellec and C?eline Rouveirol, editors, Proc. of the
10th European Conference on Machine Learning, number 1398, pages 137?142, 1998.
[2] Robert E. Schapire and Yoram Singer. BoosTexter: A boosting-based system for text
categorization. Machine Learning, 39(2/3):135?168, 2000.
[3] Naonori Ueda and Kazumi Saito. Parametoric mixture models for multi-topic text. In
Advances in Neural Information Processing Systems 15, pages 1261?1268, 2003.
[4] Koby Crammer and Yoram Singer. On the algorithmic implementation of multiclass
kernel-based vector machines. Journal of Machine Learning Research, 2:265?292,
2001.
[5] Klaus-Robert M?uller, Sebastian Mika, Gunnar R?atsch, Koji Tsuda, and Bernhard
Sch?olkopf. An introduction to kernel-based learning algorithms. IEEE Transactions
on Neural Networks, 12(2):181?201, 2001.
[6] Vladimir N. Vapnik. Statistical Learning Theory. John Wiley & Sons, Inc., 1998.
[7] Ricardo Baeza-Yates and Berthier Ribeiro-Neto. Modern Information Retrieval.
Addison-Wealy, 1999.
| 2669 |@word middle:2 polynomial:1 norm:1 km:2 decomposition:2 reduction:2 configuration:1 contains:1 hereafter:1 document:2 outperforms:1 existing:4 com:1 si:2 written:3 john:1 numerical:1 berthier:1 v:2 selected:1 directory:1 boosting:1 location:1 five:2 rc:2 along:1 combine:1 frequently:1 multi:22 kazumi:2 voc:2 mmls:3 becomes:3 notation:2 interpreted:1 corporation:2 classifier:18 k2:2 wrong:1 brute:1 unit:1 omit:1 t1:1 positive:1 treat:2 sd:4 limit:2 mika:1 chose:1 co:3 averaged:2 saito:2 dice:5 convenient:1 word:4 performace:1 cannot:1 close:2 impossible:3 www:1 map:3 independently:1 penetrate:1 regarded:1 orthonormal:1 spanned:2 imagine:1 exact:10 nippon:2 humanity:1 hypothesis:1 wolfe:1 labeled:1 observed:2 pmm:9 trained:5 solving:2 rewrite:1 learner:3 basis:1 model2:1 easily:1 mh:1 derivation:1 sc:6 labeling:24 klaus:1 whose:6 quite:2 larger:1 solve:2 say:2 s:2 otherwise:2 unseen:4 hyperlinked:1 propose:3 mli:1 maximal:11 product:6 fr:3 macro:1 relevant:7 rapidly:1 achieve:1 boostexter:4 frobenius:1 olkopf:1 tpc:2 empty:1 categorization:11 object:2 help:1 ac:2 eq:33 predicted:3 c:6 involves:1 correct:1 mtc:13 education:1 explains:1 hx:1 assign:1 generalization:2 summation:3 extension:1 pl:5 hold:1 around:1 algorithmic:1 predict:1 substituting:1 achieves:2 smallest:1 proc:1 outperformed:1 label:53 combinatorial:2 largest:1 tf:7 reflects:1 uller:1 rather:2 pn:2 joachim:1 indicates:1 mainly:1 economy:1 lj:1 issue:1 arg:6 dual:4 classification:6 among:1 yahoo:4 development:1 art:1 once:1 categorizes:1 never:1 identical:1 kw:1 represents:1 koby:1 t2:1 report:1 modern:1 randomly:2 replaced:2 taira:2 replacement:1 huge:2 evaluation:3 violation:3 mixture:4 recreation:1 behind:1 tj:1 fmeasure:1 naonori:2 partial:2 necessary:1 culture:1 eline:1 orthogonal:1 koji:1 lbl:2 tsuda:1 column:1 cover:1 ar:2 maximization:2 cost:3 too:1 dependency:1 sv:5 yuji:1 bu:2 physic:1 telegraph:2 choose:1 inefficient:1 ricardo:1 li:28 japan:1 account:1 converted:1 bold:1 inc:1 satisfy:1 caused:2 explicitly:1 lot:1 portion:1 start:1 il:2 ni:2 yield:1 weak:2 accurately:1 mc:5 explain:1 sebastian:1 ed:2 definition:1 evaluates:1 frequency:6 dataset:3 knowledge:4 appears:2 adaboost:1 done:3 evaluated:3 hand:2 web:9 scientific:1 grows:1 name:1 contain:1 normalized:1 true:5 equality:1 assigned:2 eisaku:1 laboratory:1 cosine:4 criterion:1 geometrical:1 meaning:1 image:2 novel:3 qp:8 jp:1 exponentially:1 he:2 significant:1 rd:4 seldom:2 pm:6 hxi:2 similarity:6 summands:1 closest:2 showed:1 optimizes:1 irrelevant:1 inequality:2 binary:3 underlined:3 continue:1 seen:2 multiple:1 kyoto:1 reduces:1 ntt:2 hideto:1 match:10 retrieval:12 divided:2 kaneda:1 involving:1 kernel:5 represent:2 penalize:3 separately:1 sch:1 induced:2 call:3 presence:1 baeza:1 inner:5 prototype:11 cn:2 multiclass:1 whether:1 penalty:2 speaking:1 generally:2 involve:2 listed:1 category:3 reduced:1 schapire:1 outperform:1 exist:3 judging:1 yates:1 gunnar:1 drawn:1 abstaining:1 sum:4 inverse:1 almost:1 ueda:2 decision:1 fl:2 internet:1 constraint:6 idf:3 argument:1 min:2 cslab:1 combination:3 poor:1 smaller:1 son:1 thorsten:1 computationally:2 equation:1 discus:3 count:1 singer:3 know:1 mml:22 addison:1 available:1 appropriate:1 kazawa:2 top:2 include:1 entertainment:1 yoram:2 especially:1 hikaridai:1 society:1 occurs:1 parametric:1 usual:1 md:5 distance:1 thank:1 topic:37 collected:1 code:1 ratio:8 providing:1 vladimir:1 setup:1 robert:2 statement:1 subproblems:1 negative:1 implementation:3 neto:1 datasets:7 extended:1 communication:1 community:1 pred:1 learned:1 boost:1 address:4 maeda:2 summarize:1 hyperlink:1 model1:1 rf:2 including:2 max:7 demanding:3 natural:1 force:1 circumvent:1 business:1 hm:1 health:1 text:20 prior:4 literature:2 l2:2 acknowledgement:1 relative:1 rationale:1 interesting:1 mercer:1 kecl:1 editor:1 claire:1 summary:1 placed:1 understand:1 fall:1 taking:1 regard:1 boundary:1 calculated:1 vocabulary:1 curve:1 overcome:1 quantum:6 adopts:1 made:1 collection:5 avg:1 author:1 ribeiro:1 cope:1 social:1 transaction:1 approximate:1 ignore:1 implicitly:1 bernhard:1 ml:1 assumed:1 discriminative:1 xi:9 search:1 why:2 table:12 additionally:1 european:1 whole:1 repeated:1 tl:1 en:2 wiley:1 embeds:1 sub:2 candidate:1 learns:3 ix:4 hw:1 specific:6 symbol:1 svm:5 intractable:1 burden:1 vapnik:1 sparseness:1 margin:21 easier:1 simply:1 likely:1 appearance:1 bo:5 problem2:1 goal:2 replace:1 absence:1 experimentally:1 telephone:2 except:2 total:1 called:1 experimental:1 atsch:1 rarely:1 select:2 formally:1 support:2 crammer:2 relevance:1 incorporate:1 evaluate:2 ex:2 |
1,838 | 267 | Recognizing Hand-Printed Letters and Digits
Recognizing Hand-Printed Letters and Digits
Gale L. Martin James A. Pittman
MCC, Austin, Texas 78759
ABSTRACT
We are developing a hand-printed character recognition system using a multilayered neural net trained through backpropagation. We report on results of
training nets with samples of hand-printed digits scanned off of bank checks
and hand-printed letters interactively entered into a computer through a stylus digitizer. Given a large training set, and a net with sufficient capacity to
achieve high performance on the training set, nets typically achieved error
rates of 4-5% at a 0% reject rate and 1-2% at a 10% reject rate. The topology
and capacity of the system, as measured by the number of connections in the
net, have surprisingly little effect on generalization. For those developing
practical pattern recognition systems, these results suggest that a large and
representative training sample may be the single, most important factor in
achieving high recognition accuracy. From a scientific standpoint, these results raise doubts about the relevance to backpropagation of learning models
that estimate the likelihood of high generalization from estimates of capacity.
Reducing capacity does have other benefits however, especially when the reduction is accomplished by using local receptive fields with shared weights.
In this latter case, we find the net evolves feature detectors resembling those
in visual cortex and Linsker's orientation-selective nodes.
Practical interest in hand-printed character recognition is fueled by two current technology trends: one toward systems that interpret hand-printing on hard-copy documents and one toward notebook-like computers that replace the keyboard with a stylus
digitizer. The stylus enables users to write and draw directly on a flat panel display.
In this paper, we report on results applying multi-layered neural nets trained through
backpropagation (Rumelhart, Hinton, & Williams, 1986) to both cases.
Developing pattern recognition systems is typically a two-stage process. First, intuition
and experimentation are used to select a set of features to represent the raw input pattern. Then a variety of well-developed techniques are used to optimize the classifier
system that assumes this featural representation. Most applications of backpropagation learning to character recognition use the learning capabilities only for this latter
405
406
Martin and Pittman
stage--developing the classifier system (Burr, 1986; Denker, Gardner, Graf, Henderson, Howard, Hubbard, Jackel, Baird, & Guyon, 1989; Mori & Yokosawa, 1989; Weideman, Manry, & Yau, 1989). However, backpropagation learning affords the opportunity to optimize feature selection and pattern classification simultaneously. We avoid
using pre-determined features as input to the net in favor of using a pre- segmented,
size-normalized grayscale array for each character. This is a first step toward the goal
of approximating the raw input projected onto the human retina, in that no pre-processing of the input is required.
We report on results for both hand-printed digits and letters. The hand-printed digits
come from a set of 40,000 hand-printed digits scanned from the numeric amount region
of "real-world" bank checks. They were pre-segmented and size-normalized to a
15x24 grayscale array. The test set consists of 4,000 samples and training sets varied
from 100 to 35,200 samples. Although it is always difficult to compare recognition rates
arising from different pattern sets, some appreciation for the difficulty of categorization can be gained using human performance data as a benchmark. An independent
person categorizing the test set of pre-segmented, size-normalized digits achieved an
error rate of 3.4%. This figure is considerably below the near-perfect performance of
operators keying in numbers directly from bank checks, because the segmentation algorithm is flawed .
Working with letters, as well as digits, enables tests of the generality of results on a
different pattern set having more than double the number of output categories. The
hand-printed letters come from a set of 8,600 upper-case letters collected from over
110 people writing with a stylus input device on a flat panel display. The stylUS collects
a sequence of x-y coordinates at 200 points per second at a spatial resolution of 1000
points per inch. The temporal sequence for each character is first converted to a sizenormalized bitmap array, keeping aspect ratio constant. We have found that recognition accuracy is significantly improved if these bitmaps are blurred through convolution
with a gaussian distnbution. Each pattern is represented as a 15x24 grayscale image.
A test set of 2,368 samples was extracted by selecting samples from 18 people, so that
training sets were generated by people different from those generating the test set.
Training set sizes ranged from 500 to roughly 6,300 samples.
1 HIGH RECOGNITION ACCURACY
We find relatively high recognition accuracy for both pattern sets. Thble 11 reports
the minimal error rates achieved on the test samples for both pattern sets, at various
reject rates. In the case of the hand-printed digits, the 4% error rate (0% rejects) ap1. Eff~cts of the number.of training samples and network capacity and topology are reported in the
next sectIon. Nets were tramed to error rates of 2-3%. 1i"aining began with a learning rate of .05 and
a mome~tum value of .9. The learning rate was decreased when training accuracy began to oscillate or
had stabtlized for a large number of training epochs. We evaluate the output vector on a winner-takeall basis, as this consistently improves accuracy and results in network parameters having a smaller
effect on perfonnance.
Recognizing Hand-Printed Letters and Digits
proaches the 3.4% errors made by the human judge. This suggests that further improvements to generalization will require improving segmentation accuracy. The fact
that an error rate of 5% was achieved for letters is promising. Accuracy is fairly high,
Table 1:
Error rates of best nets trained on largest sample sets and tested
on new samples
REJECT RATE
0%
5%
10%
35%
DIGITS
LETTERS
4%
5%
3%
2%
.003%
3%
1%
.001%
even though there are a large number of categories (26). This error rate may be adequate for applications where contextual constraints can be used to significantly boost
accuracy at the word-level.
2 MINIMAL NETWORK CAPACITY AND TOPOLOGY EFFECTS
The effects of network parameters on generalization have both practical and scientific
significance. The practical developer of pattern recognition systems is interested in
such effects to determine whether limited resources should be spent on trying to optimize network parameters or on collecting a larger, more representative training set.
For the scientist, effects of capacity bear on the relevance of learning models to backpropagation.
A central premise of most general models of learning-by-example is that the size of
the initial search space-the capacity of the system-determines the number of training samples needed to achieve high generalization performance. Learning is conceptualized as a search for a function that maps all possible inputs to their correct outputs.
Learning occurs by comparing successive samples of input-output pairs to functions
in a search space. Functions inconsistent with training samples are rejected. Very large
training sets narrow the search down to a function that closely approximates the desired function and yields high generalization. The capacity of a learning system-the
number of functions it can represent--determines generalization, since a larger initial
search space requires more training samples to narrow the search sufficiently . This
suggests that to improve generalization, capacity should be minimized. Unfortunately,
it is typically unclear how to minimize capacity without eliminating the desired function
from the search space. A heuristic, which is often suggested, is that simple is usually
better. It receives support from experience in curve fitting. Low-order polynomials typically extrapolate and interpolate better than high-order polynomials (Duda & Hart,
1973).
Extensions of the heuristic to neural net learning propose reducing capacity by reducing the number of connections or the number of bits used to represent each connection
407
408
Martin and Pittman
weight (Baum & Haussler, 1989; Denker, Schwartz, Wittner, Solla, Howard, Jackel,
& Hopfield,1987). We manipulated the capacity of nets in a number of ways: 1) varying
the number of hidden nodes, 2) limiting connectivity between layers so that nodes received input from only local areas, and 3) sharing connection weights between hidden
nodes. We found only negligible effects on generalization.
2.1 NUMBER OF HIDDEN NODES
Figure 1 presents generalization results as a function of training set size for nets having
one hidden layer and varying numbers of hidden nodes. The number of free parameters
(i.e., number of connections and biases) in each case is presented in parentheses. Despite considerable variation in the number of free parameters, using nets with fewer
hidden nodes did not improve generalization.
Baum & Haussler (1989) estimate the number of training samples required to achieve
an error rate e (where 0 < e ~ 1/8) on the generalization test, when an error rate
of el2 has been achieved on the training set. They assume a feed-forward net with
one hidden layer and W connections. The estimates are distribution-free in the sense
that calculations assume an arbitrary to-be-learned function. If the number of training
samples is of order : log ~ ,where N refers to the number of nodes, then it is a near
certainty that the net will achieve generalization rates of (1 - e). This estimate is the
upper-bound on the number of training samples needed. They also provide a lower
Digits
Letters
100
100
,
./
.;/
,';/
I
~
-
I
,
~
t:::
0
u
~
Number of Hidden Nodes
50
100
I
75
75
1000
I
/
,
Number of Hidden Nodes
50
(18,560)
170
(63,080)
170
(65,816)
383 (142,103)
365
(141,281)
10000 100000
Training Set Size
50
100
1000
10000 100000
'll'aining Set Size
Figure 1. Effect of number of hidden nodes and training set size on generalization.
Recognizing Hand-Printed Letters and Digits
bound estimate, on the order of W/e. Using fewer than this number of samples will,
for some functions, fail to achieve generalization rates of (1- e). The fact that we find
no advantage to reducing the number of connections conflicts with Baum & Haussler's
estimates and the underlying assumption that capacity plays a strong role in determining generalization.
Baum & Haussler also suggest using a constant of proportionality of 1 in their estimates. This implies that achieving error rates of 10% or less on new samples requires
about 10 times as many training examples as there are connection weights in the net.
For our largest nets, this implies a requirement of roughly a million training samples,
which most developers would regard as prohibitively large. We found that about 5,000
samples were sufficient. Thus, a sufficiently large training sample does not imply aprohibitively large sample, at least for character recognition. We find that sample sizes of
the order of thousands to tens of thousands yield performance very close to human levels. One reason for the discrepancy is that Baum & Haussler'S estimates are distribution-free in the sense that they reflect worst-case scenarios across all possible functions the net might learn. Presumably, the functions underlying most natural pattern
recognition tasks are not representative of the set of all possible functions. These results raise doubts about the relevance to natural pattern recognition of learning models
based on worst-case analyses, because content may greatly impact generalization.
2.2 LOCAL CONNECTMTY AND SHARED WEIGHTS
A more biologically plausible way to reduce capacity is to limit connectivity between
layers to local areas and to use shared weights. For example, visual cortex contains
neurons, each of which is responsive to a feature such as an oriented line appearing
in a small, local region on the retina (Hubel & Wiesel, 1979). A given oriented line-detector is essentially replicated across the visual field, so that the same feature can be
detected wherever it appears. In this sense, the connections feeding into an orientedline detector are shared across all similar line-detectors for different areas of the visual
field. In an artificial neural net, local structure is achieved by limiting connectivity.
A given hidden node receives input from only local areas in the input or hidden layer
preceding it. Weight sharing is achieved by linking the incoming weights across a set
of hidden nodes. Corresponding weights leading into these nodes are randomly initialized to the same values and forced to have equivalent updates during learning. In this
way the net evolves the same local feature set that is invariant across the input array.
Several demonstrations exist indicating that local connectivity and shared weights improve generalization performance in tasks where position invariance is required (Ie
Cun, 1989; Rumelhart, Hinton, & Williams, 1986).
We examined the benefits of using local receptive fields with shared weights for handprinted character recognition, where position invariance was not required. This does
not minimize the importance of position invariance. However, it is only one of many
necessary invariants underlying reliable pattern recognition. Unfortunately, most of
these invariants have not been explicitly specified. So we don't know how to bias a net
toward discovering them. Testing the role of local receptive fields with shared weights
409
410
Martin and Pittman
in situations where position invariance is not required is relevant to discovering whether these constraints have a role other than in promoting position invariance.
As indicated in Figure 2, we find only slightly improved generalization in moving from
nets with global connectivity between layers to nets with local receptive fields or to nets
with local receptive fields and shared weights. This is true despite the fact that the
number of free parameters is substantially reduced. The positive effects that do occur
are at relatively small training set sizes. This may explain why others have reported
a greater degree of improved generalization by using local receptive fields (Honavar
& Uhr, 1988). The data reported are for networks with two hidden layers. Global nets
had 150 nodes in the first layer and 50 nodes in the second. In the Local nets, first hidden layer nodes (540) received input from 5x8local and overlapping regions (offset by
2 pixels) on the input array. Second hidden layer nodes (100) and output layer nodes
had global receptive fields. The Local. Shared nets had 540 nodes in the first hidden
layer with shared weights and, at the second hidden layer, either 102 (digits) or 180 (letters) nodes with local, overlapping, and shared receptive fields of size 4x6 on the 1st
100
Digits
100
h
Letters
,~
, ,#
. ' .,f
.....
u
Q)
t:::
0
u
75
75
~
50
100
- - Global
Local
(62.210)
(77.250)
Global (63,026)
Local,
Shared
( 4,857)
Local,
1 6
Shared (11, 4 )
1000
10000 100000
Training Set Size
Figure 2.
Local
(78,866)
50 +-~:;:::==::::;::::==:::::;::::::.
100
1000 10000 100000
Training Set Size
Effects of net capacity and topology on generalization.
hidden layer. We have experimented with a large variety of different net architectures
of this sort, varying the number of hidden nodes, the sizes and overlap of local receptive
fields, and the use of local receptive fields with and without shared weights in one or
both hidden layers. The fact that we've found little difference in generalization for two
different pattern sets across such variations in network architectures argues for the
generality of the results.
Recognizing Hand-Printed Letters and Digits
2.3
DISCUSSION
Given an architecture that enables relatively high training performance, we find only
small effects of network capacity and topology on generalization performance. A large
training set yields relatively high recognition accuracy in a robust way across most net
architectures with which we've worked. These results suggest some practical advice to
those developing hand-printed character recognition systems. IT optimizing generalization performance is the goal, it is probably better to devote limited resources to collecting a very large, representative training set than to extensive experimentation with
different net architectures. The variations in net capacity and topolOgy we've examined
do not substantially affect generalization performance for sufficiently large training
sets. Sufficiently large should be interpreted as on the order of a thousand to tens of
thousands of samples for hand-printed character recognition.
From a theoretical standpoint, the negligible effects of network capacity on generalization performance contradicts the central premise of machine learning that the size of
the initial hypothesis space determines learning performance. This challenges the relevance, to backpropagation learning, of statistical models that estimate likelihood of
high generalization performance from estimates of capacity. Due to the gradient descent nature of backpropagation learning, not all functions that can be represented will
be visited during learning. The negligible effects of capacity suggest that the number
of functions visited during learning constitutes only a very small percentage of the total
possible functions that can be represented.
There are a number of reasons for believing that capacity might impact generalization
performance in other circumstances. We regularly train only to 2-3% error rates. This
helps to avoid the possibility of overfitting the data, although we have seen no indication of this when we have trained to higher levels, as long as we use large training sets.
It is also possible that the number of connections is not a good measure of capacity.
For example, the amount of information that can be passed on by a given connection
may be a better measure than the number of connections. At this conference, Ie Cun,
Denker, Solla, Howard, & Jackel have also presented evidence that removing unimportant weights from a network may be a better way to reduce capacity. However, the
fact that generalization rates come very close to human accuracy levels, even for nets
with extremely large numbers of free parameters, suggests that general effects of net
capacity and topology are, at best, small in comparison to effects of training set size.
We don't deny that there are likely to be net topologies that push performance up to
human accuracy levels, presumably by biasing the net toward discovering the range of
invariants that underlie human pattern recognition. The problem is that only a few
of these invariants have been explicitly specified (e.g., position, size, rotation), and so
it is not possible to bias a net toward discovering the full range.
411
412
Martin and Pittman
3 ADVANTAGES OF REDUCING CAPACITY
Although reducing gross indicators of capacity may not significantly improve generalization, there are good practical and scientific reasons for doing it. A good reason to
reduce the number of connections is to speed processing. Also, using local receptive
fields with shared weights biases a net toward position invariance, and toward producing a simpler, more modular internal representation which can be replicated across a
large retina. This has important implications for developing nets that combine character segmentation with recognition.
Using local receptive fields with shared weights also offers promise for increasing our
understanding of how the net correctly classifies patterns because the number of distinct receptive fields is greatly reduced. Figure 3 depicts Hinton diagrams of local reDigits
Letters
Figure 3. Receptive fields that evolved in 1st hidden layer nodes in nets with
local receptive fields having shared weights.
ceptive fields from 1st hidden layer nodes in nets with shared weights trained on digits
or letters. Each of the eight large, gray rectangles corresponds to the receptive field
for a hidden node. The four on the left came from a net trained on digits; those on
the right from a net trained on letters. Within each ofthese eight, the black rectangles
correspond to negative weights and the white to positive weights. The size of the black
and white rectangles reflects the magnitude of the weights.
The local feature detectors that develop for both pattern sets appear to be oriented
line and edge detectors. These are similar to oriented line and edge detectors found
in visual cortex (Hubel & Wiesel, 1979) and to Linsker's (1986,1988) orientation-selective nodes, which emerge from a self-adaptive net exposed to random patterns. In
Linsker's case, the feature detectors develop as an emergent property of the principle
that the signal transformation occurring from one layer to the next should maximize
the information that output signals convey about input signals. The fact that similar
Recognizing Hand-Printed Letters and Digits
feature detectors emerge in backpropagation nets trained on "natural" patterns is interesting because there were no explicit constraints to maximize information flow between layers in the backpropagation nets and because categorization is typically viewed
as an abstraction process involving considerable loss of category-irrelevant information.
References
Baum, E. and Haussler, D. (1989) What size net gives valid generalization? in D. S.
Touretzky (Ed.) Advances in neural information processing systems I, Morgan
Kaufman.
Burr, D. J. (1986) A neural network digit recognizer. Proceedings of the 1986
International conference on systems, man and cybernetics, Atlanta, Georgia.
pp. 1621-1625.
Denker, J. S., Gardner, W. R., Graf, H. P., Henderson, D., Howard, R. E., Hubbard,
W., Jackel, L. D., Baird, H. S., and Guyon, I. (1989) Neural network recognizer
for hand-written zip code digits in D. S. Touretzky (Ed.) Advances in neural
information processing systems I, Morgan Kaufman.
Denker, J. S., Schwartz, D., Wittner, B., SolIa, S., Howard, R. E., Jackel, L. D., &
Hopfield, J. (1987) Large automatic learning, rule extraction and generalization.
Complex Systems, 1, pp. 877-933.
Duda, R. 0., and Hart, P. E. (1973) Pattern classification and scene analysis. NY:
John Wiley & Sons.
Honavar, V. and Uhr, L. (1988) Experimental results indicate that generation, local
receptive fields and global convergence improve perceptual learning in
connectionist networks. CS-TR 805. Computer Science Department, University
of Wisconsin-Madison.
Hubel, D. H. and Wiesel, T. N. (1979) Brain mechanisms of vision. Scientific
American, 241, pp. 150-162.
Ie Cun, Y. (1989) Generalization and network design strategies. Thchnical Report
CRG-TR-89-4, Department of Computer Science, University of Thronto.
Linsker, R. (1986) From basic network principles to neural architecture; Emergence
of orientation-selective cells. Proceedings of the National Academy of Sciences,
USA, 83, pp. 8390-8394.
Linsker, R. (1988) Thwards an organizing principle for a layered perceptual network
in D. Anderson (Ed.) Neural information processing systems. American Institute
of Physics.
413
414
Martin and Pittman
Mori, Y. and Yokosawa, K. (1989) Neural networks that learn to discriminate similar
kanji characters. in. D. S. Touretzky (Ed.) Advances in neural information
processing systems I, Morgan Kaufman.
Rumelhart, D. E., Hinton, O. E., & Williams, R. J. Learning internal representations
by error propagation in D. E. Rumelhart & J. L. McClelland (Editors) Parallel
distributed processing: V. 1. Cambridge, Mass.: MIT Press, 1986
Weideman, W. E., Manry, M T. & Yau, H. C. (1989) A comparison of a nearest
neighbor classifier and a neural network for numeric handprint character
recognition. IEEE International Conference on Neural Networks, Washington,
D. c., 1989.
Acknowledgements
We would like to thank the NCR corporation for loaning us the set of hand-printed
digits and Joyce Conner, Janet Kilgore, and Kay Bauer for their invaluable help in collecting the set of hand-printed letters.
| 267 |@word eliminating:1 wiesel:3 thchnical:1 polynomial:2 duda:2 proportionality:1 tr:2 reduction:1 initial:3 contains:1 selecting:1 ap1:1 document:1 bitmap:2 current:1 contextual:1 comparing:1 written:1 john:1 thble:1 enables:3 update:1 fewer:2 device:1 discovering:4 manry:2 node:26 successive:1 simpler:1 consists:1 fitting:1 combine:1 burr:2 roughly:2 multi:1 brain:1 little:2 increasing:1 classifies:1 underlying:3 panel:2 mass:1 what:1 evolved:1 kaufman:3 interpreted:1 substantially:2 developer:2 developed:1 transformation:1 corporation:1 temporal:1 certainty:1 collecting:3 stylus:5 prohibitively:1 classifier:3 schwartz:2 underlie:1 digitizer:2 appear:1 producing:1 positive:2 negligible:3 scientist:1 local:29 el2:1 limit:1 despite:2 might:2 black:2 examined:2 collect:1 suggests:3 limited:2 range:2 practical:6 testing:1 backpropagation:10 digit:22 area:4 mcc:1 reject:5 printed:19 significantly:3 pre:5 word:1 refers:1 suggest:4 onto:1 close:2 layered:2 selection:1 operator:1 janet:1 applying:1 writing:1 optimize:3 equivalent:1 map:1 conceptualized:1 resembling:1 williams:3 baum:6 resolution:1 rule:1 haussler:6 array:5 kay:1 coordinate:1 variation:3 limiting:2 play:1 user:1 hypothesis:1 trend:1 rumelhart:4 recognition:22 role:3 worst:2 thousand:4 region:3 solla:2 gross:1 intuition:1 kanji:1 trained:8 raise:2 exposed:1 basis:1 hopfield:2 emergent:1 represented:3 various:1 train:1 forced:1 distinct:1 detected:1 artificial:1 heuristic:2 larger:2 plausible:1 modular:1 favor:1 emergence:1 sequence:2 advantage:2 indication:1 net:49 propose:1 relevant:1 entered:1 organizing:1 achieve:5 academy:1 proaches:1 convergence:1 double:1 requirement:1 categorization:2 perfect:1 generating:1 spent:1 help:2 develop:2 measured:1 nearest:1 received:2 strong:1 c:1 come:3 judge:1 implies:2 indicate:1 conner:1 closely:1 correct:1 uhr:2 human:7 eff:1 require:1 premise:2 feeding:1 generalization:33 crg:1 extension:1 sufficiently:4 presumably:2 notebook:1 recognizer:2 visited:2 jackel:5 hubbard:2 largest:2 reflects:1 mit:1 always:1 gaussian:1 avoid:2 varying:3 categorizing:1 improvement:1 consistently:1 check:3 likelihood:2 believing:1 greatly:2 sizenormalized:1 sense:3 abstraction:1 typically:5 hidden:24 selective:3 interested:1 pixel:1 classification:2 orientation:3 spatial:1 fairly:1 field:20 having:4 extraction:1 flawed:1 washington:1 constitutes:1 linsker:5 discrepancy:1 minimized:1 report:5 others:1 connectionist:1 few:1 retina:3 oriented:4 randomly:1 manipulated:1 simultaneously:1 ve:3 interpolate:1 national:1 atlanta:1 interest:1 possibility:1 henderson:2 implication:1 edge:2 necessary:1 experience:1 perfonnance:1 joyce:1 initialized:1 desired:2 theoretical:1 minimal:2 recognizing:6 aining:2 reported:3 considerably:1 person:1 st:3 international:2 ie:3 off:1 physic:1 connectivity:5 central:2 reflect:1 interactively:1 pittman:6 gale:1 yau:2 american:2 leading:1 doubt:2 converted:1 baird:2 blurred:1 explicitly:2 doing:1 sort:1 capability:1 parallel:1 minimize:2 accuracy:12 yield:3 correspond:1 inch:1 raw:2 cybernetics:1 detector:9 explain:1 touretzky:3 sharing:2 ed:4 pp:4 james:1 improves:1 segmentation:3 appears:1 feed:1 tum:1 higher:1 x6:1 improved:3 though:1 generality:2 anderson:1 rejected:1 stage:2 hand:21 working:1 receives:2 overlapping:2 propagation:1 gray:1 indicated:1 scientific:4 usa:1 effect:15 normalized:3 true:1 ranged:1 white:2 ll:1 during:3 self:1 trying:1 argues:1 invaluable:1 image:1 began:2 rotation:1 winner:1 keying:1 million:1 linking:1 approximates:1 interpret:1 handprint:1 cambridge:1 automatic:1 deny:1 had:4 moving:1 cortex:3 optimizing:1 irrelevant:1 scenario:1 keyboard:1 came:1 accomplished:1 seen:1 morgan:3 greater:1 preceding:1 zip:1 determine:1 maximize:2 signal:3 full:1 segmented:3 calculation:1 offer:1 long:1 wittner:2 hart:2 parenthesis:1 impact:2 involving:1 basic:1 essentially:1 circumstance:1 vision:1 represent:3 achieved:7 cell:1 decreased:1 diagram:1 standpoint:2 probably:1 regularly:1 inconsistent:1 flow:1 near:2 variety:2 affect:1 architecture:6 topology:8 reduce:3 texas:1 whether:2 passed:1 oscillate:1 adequate:1 unimportant:1 amount:2 ten:2 category:3 solia:1 reduced:2 mcclelland:1 affords:1 exist:1 percentage:1 arising:1 per:2 correctly:1 write:1 promise:1 four:1 achieving:2 rectangle:3 letter:20 fueled:1 guyon:2 draw:1 bit:1 layer:19 ct:1 bound:2 display:2 scanned:2 occur:1 constraint:3 worked:1 scene:1 flat:2 aspect:1 speed:1 extremely:1 weideman:2 martin:6 relatively:4 department:2 developing:6 honavar:2 smaller:1 across:8 slightly:1 character:12 contradicts:1 son:1 cun:3 evolves:2 biologically:1 yokosawa:2 wherever:1 ofthese:1 invariant:5 mori:2 resource:2 fail:1 mechanism:1 needed:2 know:1 experimentation:2 takeall:1 denker:5 promoting:1 eight:2 appearing:1 responsive:1 assumes:1 opportunity:1 madison:1 especially:1 approximating:1 occurs:1 receptive:17 strategy:1 unclear:1 devote:1 gradient:1 thank:1 capacity:27 collected:1 toward:8 reason:4 code:1 ratio:1 demonstration:1 handprinted:1 difficult:1 unfortunately:2 negative:1 design:1 appreciation:1 upper:2 convolution:1 neuron:1 howard:5 benchmark:1 descent:1 situation:1 hinton:4 varied:1 arbitrary:1 pair:1 required:5 specified:2 extensive:1 connection:13 conflict:1 learned:1 narrow:2 boost:1 suggested:1 below:1 pattern:20 usually:1 biasing:1 challenge:1 reliable:1 overlap:1 difficulty:1 natural:3 indicator:1 improve:5 technology:1 imply:1 gardner:2 featural:1 epoch:1 understanding:1 acknowledgement:1 determining:1 graf:2 wisconsin:1 loss:1 bear:1 interesting:1 generation:1 degree:1 sufficient:2 principle:3 editor:1 bank:3 austin:1 surprisingly:1 copy:1 keeping:1 free:6 bias:4 institute:1 neighbor:1 emerge:2 benefit:2 regard:1 curve:1 distributed:1 bauer:1 numeric:2 world:1 valid:1 forward:1 made:1 adaptive:1 projected:1 replicated:2 global:6 hubel:3 incoming:1 overfitting:1 grayscale:3 don:2 search:7 why:1 table:1 promising:1 learn:2 nature:1 robust:1 improving:1 complex:1 did:1 significance:1 multilayered:1 convey:1 advice:1 representative:4 depicts:1 georgia:1 ny:1 wiley:1 position:7 explicit:1 x24:2 perceptual:2 printing:1 down:1 removing:1 offset:1 experimented:1 evidence:1 gained:1 importance:1 magnitude:1 push:1 occurring:1 likely:1 visual:5 corresponds:1 determines:3 extracted:1 goal:2 viewed:1 shared:18 replace:1 considerable:2 hard:1 content:1 man:1 determined:1 reducing:6 total:1 discriminate:1 invariance:6 experimental:1 indicating:1 select:1 ncr:1 internal:2 people:3 support:1 latter:2 relevance:4 evaluate:1 tested:1 extrapolate:1 |
1,839 | 2,670 | A Second order Cone Programming
Formulation for Classifying Missing Data
Chiranjib Bhattacharyya
Department of Computer Science and Automation
Indian Institute of Science
Bangalore, 560 012, India
[email protected]
Pannagadatta K. S.
Department of Electrical Engineering
Indian Institute of Science
Bangalore, 560 012, India
[email protected]
Alexander J. Smola
Machine Learning Program
National ICT Australia and ANU
Canberra, ACT 0200, Australia
[email protected]
Abstract
We propose a convex optimization based strategy to deal with uncertainty
in the observations of a classification problem. We assume that instead
of a sample (xi , yi ) a distribution over (xi , yi ) is specified. In particular, we derive a robust formulation when the distribution is given by a
normal distribution. It leads to Second Order Cone Programming formulation. Our method is applied to the problem of missing data, where it
outperforms direct imputation.
1
Introduction
Denote by (x, y) ? X ?Y patterns with corresponding labels. The typical machine learning
formulation only deals with the case where (x, y) are given exactly. Quite often, however,
this is not the case ? for instance in the case of missing values we may be able (using a
secondary estimation procedure) to estimate the values of the missing variables, albeit with
a certain degree of uncertainty. It is therefore only natural to take the decreased reliability of
such data into account and design estimators accordingly. What we propose in the present
paper goes beyond the traditional imputation strategy where missing values are estimated
and then used as if they had actually been observed. The key difference in what follows is
? i is drawn from a
that we will require that with high probability any (?
xi , yi ) pair, where x
distribution of possible xi , will be estimated correctly. For the sake of simplicity we limit
ourselves to the case of binary classification.
The paper is organized as follows: Section 2 introduces the problem of classification with
uncertain data. We solve the equations arising in the context of normal random variables
in Section 3 which leads to a Second Order Cone Program (SOCP). As an application
the problem of classification with missing variables is described in Section 4. We report
experimental results in Section 5.
2
Linear Classification using Convex Optimization
Assume we have m observations (xi , yi ) drawn iid (independently and identically distributed) from a distribution over X ? Y, where X is the set of patterns and Y = {?1} are
the labels (e.g. the absence/presence of a particular object). It is our goal to find a function
f : X ? Y which classifies observations x into classes +1 and ?1.
2.1
Classification with Certainty
Assume that X is a dot product space and f is a linear function
f (x) = sgn(hw, xi + b).
(1)
In the case of linearly separable datasets we can find (w, b) which separates the two classes.
Unfortunately, such separation is not always possible and we need to allow for slack in the
separation of the two sets. Consider the formulation
minimize
w,b,?
m
X
?i
(2a)
i=1
subject to yi (hw, xi i + b) ? 1 ? ?i , ?i ? 0, kwk ? W for all 1 ? i ? m
(2b)
It is well known that this problem minimizes an upper bound on the number of errors. The
latter occur
p whenever ?i ? 1, where ?i are the slack variables. The Euclidean norm of
kwk = hw, wi, is upper bounded by a user defined constant W . This is equivalent
to lower bounding the margin, or the separation between the two classes. The resulting
discriminant surface is called the generalized optimal hyperplane [9]. The statement of (2)
is slightly nonstandard. Typically one states the SVM optimization problem as follows [3]:
m
X
1
?i
minimize kwk2 + C
w,b,?
2
i=1
(3a)
subject to yi (hw, xi i + b) ? 1 ? ?i , ?i ? 0 for all 1 ? i ? m
(3b)
Instead of the user defined parameter W , the formulation (3) uses another parameter C.
For a proper choice of C and W the two formulations are equivalent. For the purpose of
the present paper, however, (2) will be much more easily amenable to modifications and to
cast the resulting problem as a second order cone program (SOCP).
2.2
Classification with Uncertainty
So far we assumed that the (xi , yi ) pairs are known with certainty. We now relax this to the
assumption that we only have a distribution over the xi , that is (Pi , yi ) at our disposition
(due to a sampling procedure, missing variables, etc.). Formally xi ? Pi . In this case it
makes sense to replace the constraints (2b) of the optimization problem (2) by
subject to Pr {yi (hw, xi i + b) ? 1 ? ?i } ? ?i , ?i ? 0, kwk ? W ? 1 ? i ? m
(4)
Here we replaced the linear classification constraint by a probabilistic one, which is required to hold with probability ?i ? (0, 1]. This means that by choosing a value of ?i close
to 1 we can find a conservative classifier which will classify even very infrequent (x i , yi )
pairs correctly. Hence ?i provides robustness of the estimate with respect to deviating xi .
It is clear that
Pmunless we impose further restrictions on Pi , it will be difficult to minimize the
objective i=1 ?i with the constraints (4) efficiently. In the following we will consider the
special cases of gaussian uncertainty for which a mathematical programming formulation
can be found.
3
Normal Distributions
For the purpose of this section we assume that Pi = N (?
xi , ?i ), i.e., xi is drawn from a
Gaussian distribution with mean x
?i and covariance ?i . We will not require that ?i has full
rank. This means that the uncertainty about xi may be limited to individual coordinates or
to a subspace of X . As we shall see, this problem can be posed as SOCP.
3.1
Robust Classification
Under the above assumptions, the probabilistic constraint (4) becomes
subject to Pr {yi (hw, xi i + b) ? 1 ? ?i } ? ?i where xi ? N (?
x i , ?i )
?i ? 0, kwk ? W for all 1 ? i ? m
The stochastic constraint can be restated as a deterministic optimization problem
zi ? z i
yi b + ? i ? 1 ? z i
? ?i
Pr
?
? zi
? zi
(5a)
(5b)
(6)
where zi := ?yi w> xi is a normal random variable with mean z?i and variance ?z2i :=
w> ?i w. Consequently (zi ? z?i )/?zi is a random variable with zero mean and unit variance
and we can compute the lhs of (6) by evaluating the cumulative distribution function for
normal distributions
Z u
s2
1
?(u) := ?
e? 2 ds.
2? ??
In summary, (6) is equivalent to the condition
yi b + ? i ? 1 ? z i
?
? ?i .
? zi
which can be solved (since ?(u) is monotonic and invertible), for the argument of ? and
obtain a condition on its argument
p
(7)
yi (w> x
?i + b) ? 1 ? ?i + ?i wT ?i w , ?i = ??1 (?i )
We now proceed to deriving a mathematical programming formulation.
3.2
Second Order Cone Programming Formulation
Depending on ?i we can distinguish between three different cases. First consider the case
where ?i = 0 or ?i = 0.5. This means that the second order cone part of the constraint (7)
reduces to the linear inequality of (2b). In other words, we recover the linear constraint of
a standard SVM.
Secondly consider the case ?i < 0 or ?i < 0.5. This means that the constraint (7) describes
a concave set, which turns the linear classification task into a hard optimization problem.
However, it is not very likely that anyone would like to impose such constraints which hold
only with low probability. After all, uncertain data requires the constraint to become more
restrictive in holding not only for a guaranteed point xi but rather for an entire set.
Lastly consider the case ?i > 0 or ?i > 0.5 second order cone constraint. In this case (7)
describes a convex set in in w, b, ?i . We obtain the following optimization problem:
m
X
minimize
?i
(8a)
w,b,?
i=1
1
subject to yi (w> xi + b) ? 1 ? ?i + ?i k?i2 wk and ?i ? 0 ? 1 ? i ? m
kwk ? W
(8b)
(8c)
These problems can be solved efficiently by publicly available codes: recent advances in
Interior point methods for convex nonlinear optimization [8] have made such problems
feasible. As a special case of convex nonlinear optimization SOCPs have gained much
attention in recent times. For a further discussion of efficient algorithms and applications
of SOCP see [6].
3.3
Worst Case Prediction
Note that if at optimality ?i > 0, the hyperplane intersects with the constraint set
B(xi , ?i , ?i ). Moreover, at a later stage we will need to predict the class label to asses
on which side of the hyperplane B lies. If the hyperplane intersects B we will end up with
different predictions for points in the different half spaces. In such a scenario a worst case
prediction, y can be
hw, xi i + b
y = sgn(z) sgn(h ? ?) where ? = ??1 (?), z = ?
and h = |z|.
(9)
w> ?w
Here sgn(z) gives us the sign of the point in the center of the ellipsoid and (h ? ?) is
the distance of z from the center. If the hyperplane intersects the ellipsoid, the worst case
prediction is then the prediction for all points which are in the opposite half space of the
center (xi ). Plugging ? = 0.5, i.e., ? = 0 into (9) yields the standard prediction (1).
In such a case h can serve as a measure of confidence as to how well the discriminating
hyperplane classifies the mean(xi ) correctly.
3.4
Set Constraints
The same problem as (8) can also be obtained by considering that the uncertainty in each
datapoint is characterized by an ellipsoid
2
B(xi , ?i , ?i ) = {x : (x ? xi )> ??1
(10)
i (x ? xi ) ? ?i }
in conjunction with the constraint
yi (hw, xi + b) ? 1 ? ?i for all x ? Si
(11)
where Si = B(xi , ?i , ?i ) As before ?i = ??1 (?i ) for ?i ? 0. In other words, we have
?i = 0 only when the hyperplane w> x + b = 0 does not intersect the ball B(xi , ?i , ?i ).
Note that this puts our optimization setting into the same category as the knowledge-based
SVM, and SDP for invariances as all three deal with the above type of constraint (11).
More to the point, in [5] Si = S(xi , ?) is a polynomial in ? which describes the set
of invariance transforms of xi (such as distortion or translation). [4] define Si to be a
polyhedral ?knowledge? set, specified by the intersection of linear constraints.
Such considerations suggest yet another optimization setting: instead of specifying a polyhedral set Si by constraints we can also specify it by its vertices. In particular, we may set
Si to be the convex hull of a set as in Si = co{xij for 1 ? j ? mi }. By the convexity of
the constraint set itself it follows that a necessary and sufficient condition for (11) to hold
is that the inequality holds for all x ? {xij for 1 ? j ? mi }. Consequently we can replace
(11) by yi (hw, xij i + b) ? 1 ? ?i Note that the index ranges over j rather than i. Such a
setting allows us to deal with uncertainties, e.g. regarding the range of variables, which are
just given by interval boundaries, etc. The table below summarizes the five cases:
Name
Plain SVM[3]
Knowledge Based SVM[4]
Invariances [5]
Normal Distribution
Convex Hull
Set Si
{xi }
Polyhedral set
trajectory of polynomial
B(xi , ?i , ?i )
co{xij ? 1 ? j ? mi }
Optimization Problem
Quadratic Program
Quadratic Program
Semidefinite Program
Second Order Cone Program
Quadratic Program
Clearly all the above constraints can be mixed and matched and it is likely that there will be
more additions to this table in the future. More central is the notion of stating the problems
via (11) as a starting point.
4
Missing Variables
In this section we discuss how to address the missing value problem. Key is how to obtain
estimates of the uncertainty in the missing variables. Since our optimization setting allows
for uncertainty in terms of a normal distribution we attempt to estimate the latter directly.
In other words, we assume that x|y is jointly normal with mean ?y and covariance ?y .
Hence we have the following two-stage procedure to deal with missing variables:
? Estimate ?y , ?y from incomplete data, e.g. by means of the EM algorithm.
? Use the conditionally normal estimates of xmissing |(xobserved , y) in the optimization problem. This can then be cast in terms of a SOCP as described in the previous
section.
Note that there is nothing to prevent us from using other estimates of uncertainty and use
e.g. the polyhedral constraints subsequently. However, for the sake of simplicity we focus
on normal distributions in this paper.
4.1
Estimation of the model parameters
We now detail the computation of the mean and covariance matrices for the datapoints
which have missing values. We just sketch the results, for a detailed derivation see e.g. [7].
Let x ? Rd , where xa ? Rda be the vector whose values are known, while xm ? Rd?da
be the vector consisting of missing variables. Assuming a jointly normal distribution in x
with mean ? and covariance ? it follows that
>
?1
xm |xa ? N (?m + ?am ??1
aa (xa ? ?a ), ?mm ? ?am ?aa ?am ).
Here we decomposed ?, ? according to (xa , xm ) into
?aa
? = (?a , ?m ) and ? =
?>
am
?am
?mm
.
(12)
(13)
Hence, knowing ?, ? we can estimate the missing variables and determine their degree of
uncertainty. One can show that [7] to obtain ?, ? the EM algorithm reads as follows:
1. Initialize ?, ?.
2. Estimate xm |xa for all observations using (12).
3. Recompute ?, ? using the completed data set and go to step 2.
4.2
Robust formulation for missing values
As stated above, we model the missing variables as Gaussian random variables, with its
mean and covariance given by the model described in the previous section. The standard
practice for imputation is to discard the covariance and treat the problem as a deterministic
problem, using the mean as surrogate. But using the robust formulation (8) one can as well
account for the covariance.
Let ma be number of datapoints for which all the values are available, while mm be the
number of datapoints containing missing values. Then the final optimization problem reads
as follows:
minimize
w,b,?
m
X
?i
(14)
i=1
subject to yi (hw, xi i + b) ? 1 ? ?i
1
2
yj (w> xj + b) ? 1 ? ?j + ??1 (?j )k?j wk
?i ? 0
kwk ? W
?1 ? i ? ma
?ma + 1 ? j ? ma + mm
?1 ? i ? ma + mm
The mean xj has two components; xaj has values available, while the imputed vector is
given by x
?mj , via (12). The matrix ?j has all entries zero except those involving the
missing values, given by Cj , computed via (12).
The formulation (14) is an optimization problem which involves minimizing a linear objective over linear and second order cone constraints. At optimality the values of w, b,
can be used to define a classifier (1). The resulting discriminator can be used to predict
the the class label of a test datapoint having missing variables by a process of conditional
imputation as follows.
Perform the imputation process assuming that the datapoint comes from class 1(class with
label y = 1). Specifically compute the mean and covariance, as outlined in section 4.1,
and denote them by ?1 and ?1 (see (13)) respectively. The training dataset of class 1 is to
be used in the computation of ?1 and ?1 . Using the estimated ?1 and ?1 compute h as
defined in (9), and denote it by h1 . Compute the label of ?1 with the rule (1), call it y1 .
Assuming that the test data comes from class 2 (with label y = ?1) redo the entire process
and denote the resulting mean, covariance, and h by ?2 , ?2 , h2 respectively. Denote by y2
the label of ?2 as predicted by (1). We decide that the observation belongs to class with
label y? as
y? = y2 if h1 < h2 and y? = y1 otherwise
(15)
The above rule chooses the prediction with higher h value or in other words the classifier
chooses the prediction about which it is more confident. Using y? , h1 , h2 as in (15), the
worst case prediction rule (9) can be modified as follows
y = y? sgn(h ? ?) where ? = ??1 (?) and h = max(h1 , h2 )
(16)
It is our hypothesis that the formulation (14) along with this decision rule is robust to
uncertainty in the data.
5
Experiments with the Robust formulation for missing values
Experiments were conducted to evaluate the proposed formulation (14), against the standard imputation strategy. The experiment methodology consisted of creating a dataset of
missing values from a completely specified dataset. The robust formulation (14) was used
to learn a classifier on the dataset having missing values. The resulting classifier was used
to give a worst case prediction (16), on the test data. Average number of disagreements was
taken as the error measure. In the following we describe the methodology in more detail.
Consider a fully specified dataset, D = {(xi , yi )|xi ? Rd , yi ? {?1}1 ? i ? N } having
N observations, each observation is a d dimensional vector (xi ) and labels yi . A certain
fraction(f ) of the observations were randomly chosen. For each of the chosen datapoints
dm (= 0.5d) entries were randomly deleted. This then creates a dataset having N datapoints
out of which Nm (= f N, 0 ? f ? 1) of them have missing values. This data is then
randomly partitioned into test set and training set in the ratio 1 : 9 respectively. We do this
exercise to generate 10 different datasets and all our results are averaged over them.
Assuming that the conditional probability distribution of the missing variables given the
? j ) can be estimated by the
other variables is a gaussian, the mean(xj ) and the covariance (C
methods described in (4.1). The robust optimization problem was then solved for different
values of ?. The parameter ?j (= ?) is set to the same value for all the Nm datapoints. For
each value of ? the worst case error is recorded.
Experimental results are reported for three public domain datasets downloaded from uci
repository ([2]). Pima(N = 768, d = 8), Heart ( N = 270, d = 13), and Ionosphere(N =
351, d = 34), were used for experiments.
Setting ? = 0.5, yields the generalized optimal hyperplane formulation, (2). The generalized optimal hyperplane will be referred to as the nominal classifier. The nominal classifier
considers the missing values are well approximated by the mean (xj ), and there is no uncertainty.
0.5
0.6
robust
nomwc
robustwc
robust
nomwc
robustwc
robust
nomwc
robustwc
0.55
0.5
0.4
0.45
0.4
0.4
0.35
0.3
0.3
0.25
0.5
0.6
0.7
0.8
0.9
1
0.2
0.5
0.41
0.6
0.7
0.8
0.9
1
0.4
0.42
0.2
0.5
0.6
0.7
0.8
0.9
1
0.6
0.7
0.8
0.9
1
0.45
robust
nomwc
robustwc
robust
nomwc
robustwc
robust
nomwc
robustwc
0.4
0.4
0.39
0.3
0.38
0.35
0.37
0.36
0.3
0.35
0.2
0.34
0.25
0.33
0.5
0.32
0.5
0.6
0.7
0.8
0.9
1
0.6
0.7
0.8
0.9
1
0.2
0.5
Figure 1: Performance of the robust programming solution for various datasets of the UCI
database. From left to right: Pima, Ionosphere, and Heart dataset. Top: small fraction
of data with missing variables (50%), Bottom: large number of observations with missing
variables (90%)
The experimental results are summarized by the graphs(1). The robust classifier almost
always outperforms the nominal classifier in the worst case sense (compare nomwc and
robustwc). Results are presented for low(f = 0.5), and high (f = 0.9) number of missing
values. The results show that for low number of missing values(f = 0.5) the robust classifier is marginally better than the nominal classifier the gain but for large f = 0.9 the gain
is significant. This confirms that the imputation strategy fails for high noise.
The standard misclassification error for the robust classifier, using the standard prediction
(1), is also shown in the graph with the legend robust. As expected the robust classifier
performance does not deteriorate in the standard misclassification sense as ? is increased.
In summary the results seems to suggest that for low noise level the nominal classifier
trained on imputed data performs as good as the robust formulation. But for high noise
level the robust formulation yields dividends in the worst case sense.
6
Conclusions
An SOCP formulation was proposed for classifying noisy observations and the resulting
formulation was applied to the missing data case. In the worst case sense the classifier
shows a better performance over the standard imputation strategy. Closely related to this
work is the Total Support Vector Classification(TSVC) formulation, presented in [1]. The
TSVC formulation tries to reconstruct the original maximal margin classifier in the presence of noisy data. Both TSVC formulation and the approach in this paper address the issue
of uncertainty in input data and it would be an important research direction to compare the
two approaches.
Acknowledgements CB was partly funded by ISRO-IISc Space technology cell (Grant
number IST/ECA/CB/152). National ICT Australia is funded through the Australian Government?s Backing Australia?s Ability initiative, in part through the Australian Research
Council. AS was supported by grants of the ARC. We thank Laurent ElGhaoui, Michael
Jordan, Gunnar R?atsch, and Frederik Schaffalitzky for helpful discussions and comments.
References
[1] J. Bi and T. Zhang. Support vector classification with input data uncertainty. In Advances in Neural Information Processing Systems. MIT Press, 2004.
[2] C. L. Blake and C. J. Merz. UCI repository of machine learning databases, 1998.
[3] C. Cortes and V. Vapnik. Support vector networks. Machine Learning, 20:273?297,
1995.
[4] G. Fung, O. L. Mangasarian, and Jude Shavlik. Knowledge-based support vector machine classifiers. In Advances in Neural Information Processing Systems. MIT Press,
2002.
[5] Thore Graepel and Ralf Herbrich. Invariant pattern recognition by semidefinite programming machines. In Advances in Neural Information Processing Systems 16, Cambridge, MA, 2003. MIT Press.
[6] M.S. Lobo, L. Vandenberghe, S. Boyd, and H. Lebret. Applications of second-order
cone programming. Linear Algebra and its Applications, 284(1?3):193?228, 1998.
[7] K. V. Mardia, J. T. Kent, and J. M. Bibby. Multivariate Analysis. Academic Press,
1979.
[8] Y. Nesterov and A. Nemirovskii. Interior Point Algorithms in Convex Programming.
Number 13 in Studies in Applied Mathematics. SIAM, Philadelphia, 1993.
[9] V. Vapnik. The Nature of Statistical Learning Theory. Springer, New York, 1995.
| 2670 |@word repository:2 polynomial:2 norm:1 seems:1 confirms:1 covariance:10 kent:1 bhattacharyya:1 outperforms:2 si:8 yet:1 half:2 accordingly:1 recompute:1 provides:1 herbrich:1 zhang:1 five:1 mathematical:2 along:1 direct:1 become:1 initiative:1 polyhedral:4 deteriorate:1 expected:1 sdp:1 decomposed:1 considering:1 becomes:1 iisc:3 classifies:2 bounded:1 moreover:1 matched:1 what:2 minimizes:1 certainty:2 act:1 concave:1 exactly:1 classifier:17 unit:1 grant:2 before:1 engineering:1 treat:1 limit:1 laurent:1 xaj:1 au:1 specifying:1 co:2 limited:1 range:2 bi:1 averaged:1 yj:1 practice:1 procedure:3 intersect:1 boyd:1 word:4 confidence:1 suggest:2 close:1 interior:2 put:1 context:1 restriction:1 equivalent:3 deterministic:2 missing:30 center:3 go:2 attention:1 starting:1 independently:1 convex:8 restated:1 simplicity:2 estimator:1 rule:4 deriving:1 vandenberghe:1 datapoints:6 ralf:1 notion:1 coordinate:1 nominal:5 infrequent:1 user:2 programming:9 us:1 hypothesis:1 approximated:1 recognition:1 database:2 observed:1 bottom:1 electrical:1 solved:3 worst:9 convexity:1 nesterov:1 lobo:1 trained:1 algebra:1 serve:1 creates:1 completely:1 easily:1 various:1 intersects:3 derivation:1 describe:1 choosing:1 quite:1 whose:1 posed:1 solve:1 z2i:1 distortion:1 relax:1 otherwise:1 reconstruct:1 ability:1 jointly:2 itself:1 noisy:2 final:1 propose:2 product:1 maximal:1 uci:3 object:1 derive:1 depending:1 stating:1 predicted:1 involves:1 come:2 australian:2 direction:1 closely:1 stochastic:1 hull:2 subsequently:1 australia:4 sgn:5 public:1 require:2 government:1 rda:1 frederik:1 secondly:1 hold:4 mm:5 blake:1 normal:11 cb:2 predict:2 purpose:2 estimation:2 label:10 council:1 mit:3 clearly:1 always:2 gaussian:4 modified:1 rather:2 conjunction:1 focus:1 rank:1 sense:5 am:5 helpful:1 typically:1 entire:2 backing:1 issue:1 classification:12 special:2 ernet:2 initialize:1 having:4 sampling:1 future:1 report:1 bangalore:2 randomly:3 national:2 individual:1 deviating:1 replaced:1 ourselves:1 consisting:1 attempt:1 introduces:1 semidefinite:2 amenable:1 necessary:1 lh:1 dividend:1 incomplete:1 euclidean:1 uncertain:2 instance:1 classify:1 increased:1 bibby:1 disposition:1 vertex:1 entry:2 conducted:1 reported:1 chooses:2 confident:1 siam:1 discriminating:1 probabilistic:2 invertible:1 michael:1 central:1 nm:2 recorded:1 containing:1 creating:1 account:2 socp:6 summarized:1 wk:2 automation:1 later:1 h1:4 try:1 kwk:6 socps:1 recover:1 minimize:5 ass:1 publicly:1 variance:2 efficiently:2 yield:3 iid:1 marginally:1 trajectory:1 nonstandard:1 datapoint:3 whenever:1 against:1 dm:1 mi:3 gain:2 dataset:7 knowledge:4 organized:1 cj:1 graepel:1 actually:1 higher:1 methodology:2 specify:1 formulation:25 just:2 smola:2 lastly:1 stage:2 xa:5 d:1 sketch:1 nonlinear:2 thore:1 name:1 consisted:1 y2:2 hence:3 read:2 i2:1 deal:5 conditionally:1 generalized:3 eca:1 performs:1 consideration:1 mangasarian:1 kwk2:1 significant:1 cambridge:1 rd:3 outlined:1 mathematics:1 had:1 funded:2 reliability:1 dot:1 surface:1 etc:2 multivariate:1 recent:2 belongs:1 discard:1 scenario:1 certain:2 inequality:2 binary:1 yi:22 impose:2 determine:1 full:1 reduces:1 characterized:1 academic:1 plugging:1 prediction:11 involving:1 jude:1 cell:1 addition:1 decreased:1 interval:1 comment:1 subject:6 legend:1 jordan:1 call:1 ee:1 presence:2 identically:1 pannaga:1 xj:4 zi:7 opposite:1 regarding:1 knowing:1 redo:1 proceed:1 york:1 tsvc:3 clear:1 detailed:1 transforms:1 category:1 imputed:2 generate:1 xij:4 sign:1 estimated:4 arising:1 correctly:3 shall:1 ist:1 key:2 gunnar:1 drawn:3 imputation:8 deleted:1 prevent:1 graph:2 fraction:2 cone:10 uncertainty:15 almost:1 decide:1 separation:3 decision:1 summarizes:1 bound:1 guaranteed:1 distinguish:1 quadratic:3 occur:1 constraint:21 alex:1 sake:2 argument:2 anyone:1 optimality:2 separable:1 department:2 fung:1 according:1 ball:1 describes:3 slightly:1 em:2 wi:1 partitioned:1 modification:1 invariant:1 pr:3 taken:1 heart:2 chiranjib:1 equation:1 slack:2 turn:1 discus:1 end:1 available:3 disagreement:1 robustness:1 original:1 top:1 completed:1 restrictive:1 objective:2 strategy:5 traditional:1 surrogate:1 subspace:1 distance:1 separate:1 thank:1 considers:1 discriminant:1 assuming:4 xobserved:1 code:1 index:1 ellipsoid:3 ratio:1 minimizing:1 difficult:1 unfortunately:1 statement:1 holding:1 pima:2 stated:1 design:1 proper:1 perform:1 upper:2 observation:10 datasets:4 arc:1 nemirovskii:1 y1:2 pair:3 cast:2 specified:4 required:1 discriminator:1 lebret:1 address:2 able:1 pannagadatta:1 beyond:1 below:1 pattern:3 xm:4 elghaoui:1 program:8 max:1 misclassification:2 natural:1 technology:1 philadelphia:1 ict:2 acknowledgement:1 fully:1 mixed:1 h2:4 downloaded:1 degree:2 sufficient:1 classifying:2 pi:4 translation:1 summary:2 supported:1 side:1 allow:1 india:2 institute:2 shavlik:1 distributed:1 boundary:1 plain:1 evaluating:1 cumulative:1 made:1 far:1 schaffalitzky:1 assumed:1 xi:39 table:2 mj:1 learn:1 robust:22 nature:1 csa:1 domain:1 da:1 linearly:1 bounding:1 s2:1 noise:3 nothing:1 canberra:1 referred:1 fails:1 exercise:1 lie:1 mardia:1 hw:10 svm:5 cortes:1 ionosphere:2 albeit:1 vapnik:2 gained:1 anu:2 margin:2 intersection:1 likely:2 monotonic:1 springer:1 aa:3 ma:6 chiru:1 conditional:2 goal:1 consequently:2 replace:2 absence:1 feasible:1 hard:1 typical:1 except:1 specifically:1 hyperplane:9 wt:1 conservative:1 called:1 total:1 secondary:1 invariance:3 experimental:3 partly:1 merz:1 atsch:1 formally:1 support:4 latter:2 alexander:1 indian:2 evaluate:1 |
1,840 | 2,671 | At the Edge of Chaos: Real-time Computations and
Self-Organized Criticality in Recurrent Neural Networks
Thomas Natschl?ager
Software Competence
Center Hagenberg
A-4232 Hagenberg, Austria
[email protected]
Nils Bertschinger
Max Planck Institute for
Mathematics in the Sciences
D-04103 Leipzig, Germany
[email protected]
Robert Legenstein
Institute for Theoretical
Computer Science, TU Graz
A-8010 Graz, Austria
[email protected]
Abstract
In this paper we analyze the relationship between the computational capabilities of randomly connected networks of threshold gates in the timeseries domain and their dynamical properties. In particular we propose
a complexity measure which we find to assume its highest values near
the edge of chaos, i.e. the transition from ordered to chaotic dynamics.
Furthermore we show that the proposed complexity measure predicts the
computational capabilities very well: only near the edge of chaos are
such networks able to perform complex computations on time series. Additionally a simple synaptic scaling rule for self-organized criticality is
presented and analyzed.
1
Introduction
It has been proposed that extensive computational capabilities are achieved by systems
whose dynamics is neither chaotic nor ordered but somewhere in between order and chaos.
This has led to the idea of ?computation at the edge of chaos?. Early evidence for this
hypothesis has been reported e.g. in [1]. The results of numerous computer simulations
carried out in these studies suggested that there is a sharp transition between ordered and
chaotic dynamics. Later on this was confirmed by Derrida and others [2]. They used
ideas from statistical physics to develop an accurate mean-field theory which allowed to
determine the critical parameters analytically. Because of the physical background, this
theory focused on the autonomous dynamics of the system, i.e. its relaxation from an
initial state (the input) to some terminal state (the output) without any external influences.
In contrast to such ?off-line? computations, we will focus in this article on time-series
computations, i.e. mappings, also called filters, from a time-varying input signal to a timevarying output signal. Such ?online? or real-time computations describe more adequately
the input to output relation of systems like animals or autonomous robots which must react
in real-time to a continuously changing stream of sensory input.
The purpose of this paper is to analyze how the computational capabilities of randomly
connected recurrent neural networks in the domain of real-time processing and the type
of dynamics induced by the underlying distribution of synaptic weights are related to each
other. In particular, we will show that for the types of neural networks considered in this paper (defined in Sec. 2) there also exists a transition from ordered to chaotic dynamics. This
phase transition is determined using an extension of the mean-field approach described in
[3] and [4] (Sec. 3). As the next step we propose a novel complexity measure (Sec. 4) which
input
neuron #
network activity
10
20
30
0
20
40
0
timesteps
20
40
0
20
timesteps
40
timesteps
0.2
0.8
?
0
0.6
?0.4
?0.6
ordered
0.1
ical
crit
?0.2
1
chaotic
0.4
mean activity
0.4
10
?2
Figure 1: Networks of randomly connected threshold gates can exhibit ordered, critical and
chaotic dynamics. In the upper row examples of the temporal evolution of the network state
xt are shown (black: xi,t = 1, white: xi,t = 0, input as indicated above) for three different
networks with parameters taken from the ordered, critical and chaotic regime, respectively.
Parameters: K = 5, N = 500, u
? = ?0.5, r = 0.3 and ? and ? 2 as indicated in the phase
plot below. The background of the phase plot shows the mean activity a? (see Sec. 3) of
the networks depending on the parameters ? and ? 2 .
can be calculated using the mean-field theory developed in Sec. 3 and serves as a predictor for the computational capability of a network in the time-series domain. Employing a
recently developed framework for analyzing real-time computations [5, 6] we investigate
in Sec. 5 the relationship between network dynamics and the computational capabilities in
the time-series domain. In Sec. 6 of this paper we propose and analyze a synaptic scaling
rule for self-organized criticality (SOC) for the types of networks considered here. In contrast to previous work [7], we do not only check that the proposed rule shows adaptation
towards critical dynamics, but also show that the computational capabilities of the network
are actually increased if the rule is applied.
Relation to previous work: In [5], the so-called liquid state machine (LSM) approach was
proposed and used do analyze the computational capabilities in the time-series domain of
randomly connected networks of biologically inspired network models (composed of leaky
integrate-and-fire neurons). We will use that approach to demonstrate that only near the
edge of chaos, complex computations can be performed (see Sec. 5). A similar analysis for
a restricted case (zero mean of synaptic weights) of the network model considered in this
paper can be found in [4].
2
The Network Model and its Dynamics
We consider input driven recurrent networks consisting of N threshold gates with states
xi ? {0, 1}. Each node i receives nonzero incoming weights wij from exactly K randomly
chosen nodes j. Each nonzero connection weight wij is randomly drawn from a Gaussian
distribution with mean ? and variance ? 2 . Furthermore, the network is driven by an external input signal u(?) which is injected into each node. Hence, in summary, the update of
PN
the network state xt = (x1,t , . . . , xN,t ) is given by xi,t = ?( j=1 wij ? xj,t?1 + ut?1 )
which is applied to all neurons in parallel and where ?(h) = 1 if h ? 0 and ?(h) = 0
otherwise. In the following we consider a randomly drawn binary input signal u (?) : at each
time step ut assumes the value u
? + 1 with probability r and the value u
? with probability
1 ? r. This network model is similar to the one we have considered in [4]. However it
differs in two important aspects: a) By using states xi ? {0, 1} we emphasis the asymmetric information encoding by spikes prevalent in biological neural systems and b) it is more
general in the sense that the Gaussian distribution from which the non-zero weights are
drawn is allowed to have an arbitrary mean ? ? R. This implies that the network activity
PN
at = N1 i=1 xi,t can vary considerably for different parameters (compare Fig. 1) and
enters all the calculations discussed in the rest of the paper.
The top row of Fig. 1 shows typical examples of ordered, critical and chaotic dynamics (see
the next section for a definition of order and chaos). The system parameters corresponding
to each type of dynamics are indicated in the lower panel (phase plot). We refer to the
(phase) transition from the ordered to the chaotic regime as the critical line (shown as the
solid line in the phase plot). Note that increasing the variance ? 2 of the weights consistently
leads to chaotic behavior.
3
The Critical Line: Order and Fading Memory versus Chaos
To define the chaotic and ordered phase of an input driven network we use an approach
which is similar to that proposed by Derrida and Pomeau [2] for autonomous systems:
consider two (initial) network states with a certain (normalized) Hamming distance. These
states are mapped to their corresponding successor states (using the same weight matrix)
with the same input in each case and the change in the Hamming distance is observed. If
small distances tend to grow this is a sign of chaos whereas if the distance tends to decrease
this is a signature of order.
Following closely the arguments in [4, 3] we developed a mean-field theory (see [8] for
all details) which allows to calculate the update dt+1 = f (dt , at , ut ) of the normalized
? t as well as the
Hamming distance dt = |{i : xi,t 6= x
?i,t }|/N between two states xt and x
update at+1 = A(at , ut ) of the network activity in one time step. Note that dt+1 depends
on the input ut (in contrast to [3]) and also on the activity at (in contrast to [4]). Hence the
two-dimensional map Fu (dt , at ) := (dt+1 , at+1 ) = (f (dt , at , ut ), A(at , ut )) describes
the time evolution of dt and at given the input times series u(?) .
Let us consider the steady state of the averaged Hamming distance f ? as well as the steady
state of the averaged network activity a? , i.e. (f ? , a? ) = limt?? hFut i.1 If f ? = 0 we
know that any state differences will eventually die out and the network is in the ordered
phase. If on the other hand a small difference is amplified and never dies out we have
f ? 6= 0 and the network is in the chaotic phase. Whether f ? = 0 or f ? 6= 0 can be decided
by looking at the slope of the function f (?, ?, ?) at its fixed point f ? = 0. Since at does not
depend on dt we calculate the averaged steady state activity a? and determine the slope ??
of the map rf (d, a, u
? + 1) + (1 ? r)f (d, a, u
?) at the point (d, a) = (0, a? ). Accordingly
we say that the network is in the ordered, critical or chaotic regime if ? ? < 1, ?? = 1 or
?? > 1 respectively. In [8] it is shown that the so called critical line ?? = 1 where the
phase transition from ordered to chaotic behavior occurs is given by
Pbf =
K?1
X
n=0
K ? 1 ?n
1
a (1 ? a? )K?1?n (rQ(1, n, u
? + 1) + (1 ? r)Q(1, n, u
?)) =
(1)
n
K
where Pbf denotes the probability (averaged over the inputs and the network activity) that
a node will change its output if a single out of its K input bits is flipped.2 Examples of
Fut denotes t-fold composition of the map Fu (?, ?) where in the k-th iteration the input uk is
applied and h?i denotes the average over all possible initial conditions and all input signals with a
given statistics determined by u
? and r.
2
The actual single bit-flip probability Q depends on the number n of inputs which are 1 and the
1
K=5
K = 10
0.4
0.1
0.3
0.2
?
0
0.06
?0.1
?0.2
0.04
?0.3
0.3
0.07
0.2
0.06
0.1
0.05
0
?
0.1
NM?Separation
0.08
?0.1
0.04
?0.2
0.03
NM?Separation
0.4
?0.3
0.02
?0.4
0.02
?0.4
?0.5
?0.5
?0.6
?0.6
0
0.1
1
?2
10
0.01
0
0.1
1
?2
10
Figure 2: N M -separation assumes high values on the critical line. The gray coded image
shows the N M -separation in dependence on ? and ? 2 for K denoted in the panels, r = 0.3,
u
? = ?0.5 and b = 0.1. The solid line marks the critical values for ? and ? 2 .
critical lines that were calculated from this formula (marked by the solid lines) can be seen
in Fig. 2 for K = 5 and K = 10.3
The ordered phase can also be described by using the notion of fading memory (see [5] and
the references therein). Intuitively speaking in a network with fading memory a state x t is
fully determined by a finite history ut?T , ut?T +1 , . . . , ut?1 , ut of the input u(?) . A slight
reformulation of this property (see [6] and the references therein) shows that it is equivalent
to the requirement that all state differences vanish, i.e. being in the ordered phase. Fading
memory plays an important role in the ?liquid state machine? framework [5] since together
with the separation property (see below) it would in principle allow an appropriate readout
function to deduce the recent input, or any function of it, from the network state. If on
the other hand the network does not have fading memory (i.e. is in the chaotic regime)
a given network state xt also contains ?spurious? information about the initial conditions
and hence it is hard or even impossible to deduce any features of the recent input.
4
NM-Separation as a Predictor for Computational Power
The already mentioned separation property [5] is especially important if a network is to
be useful for computations on input time-series: only if different input signals separate
the network state, i.e. different inputs result in different states, it is possible for a readout
function to respond differently. Hence it is necessary that any two different input time
series for which the readout function should produce different outputs drive the recurrent
network into two sufficiently different states.
The mean field theory we have developed (see [8]) can be extended to describe the
update dt+1 = s(dt , ...) of the Hamming distance that result from applying different inputs u(?) and u
?(?) with a mean distance of b := Pr {ut 6= u
?t }, i.e. the separation. In summary the three-dimensional map Su,?u (dt , at , a
?t ) := (dt+1 , at+1 , a
?t+1 ) =
(s(dt , at , a
? t , ut , u
?t ), A(at , ut ), A(?
at , u
?t )) fully describes the time evolution of the Hamming distance and the network activities. Again we consider the steady state of the averaged
t
Hamming distance s? and the network activities a? , a
?? , i.e. (s? , a? , a
?? ) = limt?? Su,?
u .
The overall separation for a given input statistics (determined by u
?, r, and b) is then given
by s? . However, this overall separation measure can not be directly related to the computa-
R ?u
external input u and is given by Q(1, n, u) = ?? ?(?, n?, n? 2 ) 1 ? ?(?u ? ?, ?, ? 2 ) d? +
R?
?(?, n?, n? 2 )?(?u ? ?, ?, ? 2 )d? where ?, ? denote the Gaussian density and cumulative den?u
sity respectively (see [8] for a detailed explanation).
3
For each value of ? = ?0.6 + k ? 0.01, k = 0 . . . 100 a search was conducted to find the value
for ? 2 such that ?? = 1. Numerical iterations of the function A were used to determine a ? .
A
B
3bit parity (K = 5)
C5bit random boolean functions
3bit parity (K = 10)
0.4
5
0.8
4
0.2
?0.2
2
?0.4
?0.6
0.01
3
2
1
2
?
10
100
0
0.4
0.2
1
1
0.1
0.6
MC (MI)
?
3
MC (MI)
0
0.01
0.1
1
2
?
10
100
0
mean MI
4
0.01
0.1
1
2
?
10
100
0
Figure 3: Real-time computation at the edge of chaos. A The gray coded image (an interpolation between the data points marked with open diamonds) shows the performance
of trained networks in dependence of the parameters ? and ? 2 for the
P delayed 3-bit parity task. Performance is measured as the memory capacity M C = ? I(v, y (? ) ) where
I(v, y (? ) ) is the mutual information between the classifier output v(?) and the target func(? )
tion yt = PARITY(ut?? , ut?? ?1 , ut?? ?2 ) measured on a test set. B Same as panel A
but for K = 10. C Same as panel A but for an average over 50 randomly drawn Boolean
functions f of the last 5 time steps, i.e. yt = f (ut , ut?1 , ..., ut?4 ).
tional power since chaotic networks separate even minor differences in the input to a very
high degree. The part of this separation that is caused by the input distance b and not by the
distance of some initial state is therefore given by s? ? f ? because f ? measures the state
distance that is caused by differences in the initial states and remains even after long runs
with the same inputs (see Sec. 3). Note that f ? is always zero in the ordered phase and
non-zero in the chaotic phase.
Since we want the complexity measure, which we will call N M -separation, to be a
predictor for computational power we correct s? ? f ? by a term which accounts for
the separation due to an all-dominant input drive. A suitable measure for this ?immediate separation? i? is the average increase in the Hamming distance if the system is
run for a long time (t ? ?) with equal inputs u(?) = u
?(?) and then a single step
with an input pair (v, v?) with an average difference
of b = Pr {v, 6= v?} is applied:
P1
t
i? = limt?? v,?v=0 rv (1?r)1?v b|v??v| (1?b)1?|v??v| s(?, ?, ?, v, v?) ? Su,u
?f ? . Hence
a measure of the network mediated separation N Msep due to input differences is given by
N Msep = s? ? f ? ? i?
(2)
In Fig. 2 the N M -separation resulting from an input difference of b = 0.1 is shown in
dependence of the network parameters ? and ? 2 .4 Note that the N M -separation peaks
very close to the critical line. Because of the computational importance of the separation
property this also suggests that the computational capabilities of the networks will peak at
the onset of chaos, which is confirmed in the next section.
5
Real-Time Computations at the Edge of Chaos
To access the computational power of a network we make use of the so called ?liquid state
machine? framework which was proposed by Maass et.al. [5] and independently by Jaeger
[6]. They put forward the idea that any complex time-series computation can be implemented by composing a system which consists of two conceptually different parts: a) a
4
For each value of ? = ?0.6 + k ? 0.05, k = 0 . . . 20, 10 values for ? 2 where chosen near
the critical line and 10 other values where equally spaced (on a logarithmic scale) over the interval
[0.02,50]. For each such pair (?, ? 2 ) extensive numerical iterations of the map S where performed
to obtain accurate estimates of s? , f ? and i? . Hopefully these numerical estimates can be replaced
by analytic results in the future.
properly chosen general-purpose recurrent network with ?rich? dynamics and b) a readout function that is trained to map the network state to the desired outputs (see [5, 6, 4]
for more details). This approach is potentially successful if the general-purpose network
encodes the relevant features of the input signal in the network state in such a way that
the readout function can easily extract it. We will show that near the critical line the networks considered in this paper encode the input in such a way that a simple linear classifier
C(xt ) = ?(w ? xt + w0 ) suffices to implement a broad range of complex nonlinear filters. Note that in order to train the network for a given task only the parameters w ? R N ,
w0 ? R of the linear classifier are adjusted such that the actual network output vt = C(xt )
is as close as possible to the target values yt .
To access the computational power in a principled way networks with different parameters
were tested on a delayed 3-bit parity task for increasing delays and on randomly drawn
Boolean functions of the last 5 input bits. Note that these tasks are quite complex for
the networks considered here since most of them are not linear separable (i.e. the parity
function) and require memory. Hence to achieve good performance it is necessary that a
state xt contains information about several input bits ut0 , t0 < t in a nonlinear transformed
form such that a linear classifier C is sufficient to perform the nonlinear computations.
The results are summarized in Fig. 3 where the performance (measured in terms of mutual
information) on a test set between the network output and the target signal is shown for
various parameter settings (for details see [4]). The highest performance is clearly achieved
for parameter values close to the critical line where the phase transition occurs. This has
been noted before [1]. In contrast to these previous results the networks used here are
not optimized for any specific task but their computational capabilities are assessed by
evaluating them for many different tasks. Therefore a network that is specifically designed
for a single task will not show a good performance in this setup. These considerations
suggest the following hypotheses regarding the computational function of generic recurrent
neural circuits: to serve as a general-purpose temporal integrator, and simultaneously as a
kernel (i.e., nonlinear projection into a higher dimensional space) to facilitate subsequent
(linear) readout of information whenever it is needed.
6
Self-Organized Criticality via Synaptic Scaling
Since the computational capabilities of a network depend crucially on having almost critical
dynamics an adaptive system should be able to adjust its dynamics accordingly.
Equ. (1) states that critical dynamics are achieved if the probability Pbf that a single bitflip in the input shows up in the output should on average (over the external and internal
1
input statistics given by u
?, r and a? respectively) be equal to K
. To allow for a rule that
can adjust the weights of each node a local estimate of Pbf must be available. This can
be accomplished by estimating Pbf from the margin of each node, i.e. the distance of the
internal activation from the firing threshold. Intuitively a node with an activation that is
much higher or lower than its firing threshold is rather unlikely to change its output if a
i
single bit in its input is flipped. Formally Pbf
of node i is given by the average (over the
internal and external input statistics) of the following quantity:
1
K
N
X
? (wij (1 ? 2xj,t?1 )(1 ? 2xi,t ) ? mi,t )
(3)
j=1,wij 6=0
P
N
where mi,t = j=1 wij xj,t?1 + ut?1 denotes the margin of node i (see [8] for details).
Each node now applies synaptic scaling to adjust itself towards the critical line. Accordingly we arrive at the following SOC-rule:
1
esti
1
if Pbf
(t) > K
1+? ? wij
wij (t + 1) =
(4)
esti
1
(1 + ?) ? wij (t) if Pbf (t) < K
A
neuron #
50
100
150
200
100
200
300
400
500
600
700
timesteps
B
C
1
1
K*P
K*P
bf
1.5
bf
1.5
0.5
0
0
0.5
100
200
300
400
timesteps
500
600
700
0
0
100
200
300
400
500
600
700
timesteps
Figure 4: Self-organized criticality. A Time evolution of the network state x t starting
in a chaotic regime while the SOC-rule (4) is active (black: xi,t = 1, white: xi,t = 0).
Parameters: N = 500, K = 5, u
? = ?0.5, r = 0.3, ? = 0 and initial ? 2 = 100. B
Estimated Pbf . The dotted line shows how the node averaged estimate of Pbf evolves over
time for the network shown in A. The running average of this estimate (thick black line) as
used by the SOC-rule clearly shows that Pbf approaches its critical value (dashed line). C
Same as B but for K = 10 and initial ? 2 = 0.01 in the ordered regime.
esti
where 0 < ? 1 is the learning rate and Pbf
(t) is a running average of the formula
i
in Equ. (3) to estimate Pbf . Applying this rule in parallel to all nodes of the network is
then able to adjust the network dynamics towards criticality as shown in Fig. 4 5 . The upper
row shows the time evolution of the network states xt while the SOC-rule (4) is running.
It is clearly visible how the network dynamics changes from chaotic (the initial network
had the parameters K = 5, ? = 0 and ? 2 = 100) to critical dynamics that respect the
input signal. The lower row of Fig. 4 shows how the averaged estimated bit-flip probability
PN
esti
1
i=1 Pbf (t) approaches its critical value for the case of the above network and one
N
that started in the ordered regime (K = 10, ? = 0, ? 2 = 0.01).
Since critical dynamics are better suited for information processing (see Fig. 3) it is expected that the performance on the 3-bit parity task improves due to SOC. This is confirmed in Fig. 5 which shows how the memory capacity M C (defined in Fig. 3) grows for
networks that were initialized in the chaotic and ordered regime respectively. Note that the
performance reached by these networks using the SOC-rule (4) is as high as for networks
where the critical value for ? 2 is chosen apriori and stays at this level. This shows that
rule (4) is stable in the sense that it keeps the dynamics critical and does not destroy the
computational capabilities.
7
Discussion
We developed a mean-field theory for input-driven networks which allows to determine the
position of the transition line between ordered and chaotic dynamics with respect to the
5
Here a learning rate of ? = 0.01 and an exponentially weighted running average with a time
constant of 15 time steps were used.
A
B
K = 5, start ?2 = 100 (chaotic)
K = 10, start ?2 = 0.01 (ordered)
5
4
MC [bits]
MC [bits]
4
3
2
3
2
1
1
0
0
0
500
1000
SOC steps
1500
2000
0
500
1000
1500
2000
SOC steps
Figure 5: Time evolution of the performance with activated SOC-rule. A The plot shows
the memory capacity M C (see Fig. 3) on the 3-bit parity task averaged over 25 networks (?
standard deviation as error-bars) evaluated at the indicated time steps. At each evaluation
time step the network weights were fixed and the M C was measured as in Fig. 3 by training
the corresponding readouts from scratch. The networks were initialized in the chaotic
regime. B Same as in A but for K = 10 and networks initialized in the ordered regime.
parameters controlling the network connectivity and input statistics. Based on this theory
we proposed a complexity measure (called N M -separation) which assumes its highest
values at the critical line and shows a clear correlation with the computational power for
real-time time-series processing. These results provide further evidence for the idea of
?computation at the edge of chaos? [1] and support the hypothesis that dynamics near
the critical line are expected to be a general property of input driven dynamical systems
which support complex real-time computations. Therefore our analysis and the proposed
complexity measure provide a new approach towards discovering dynamical principles that
enable biological systems to do sophisticated information processing.
Furthermore we have shown that a local rule for synaptic scaling is able to adjust the
weights of a network towards critical dynamics. Additionally networks adjusting themselves by this rule have been found to exhibit enhanced computational capabilities. Thereby
systems can combine task-specific optimization provided by (supervised) learning rules
with self-organization of its dynamics towards criticality. This provides an explanation
how specific information can be processed while still being able to react to incoming signals in a flexible way.
Acknowledgement This work was supported in part by the PASCAL project #IST-2002-506778
of the European Community.
References
[1] C. G. Langton. Computation at the edge of chaos. Physica D, 42, 1990.
[2] B. Derrida and Y. Pomeau. Random networks of automata: A simple annealed approximation.
Europhys. Lett., 1:45?52, 1986.
[3] B. Derrida. Dynamical phase transition in non-symmetric spin glasses. J. Phys. A: Math. Gen.,
20:721?725, 1987.
[4] N. Bertschinger and T. Natschl?ager. Real-time computation at the edge of chaos in recurrent
neural networks. Neural Computation, 16(7):1413?1436, 2004.
[5] W. Maass, T. Natschl?ager, and H. Markram. Real-time computing without stable states: A new
framework for neural computation based on perturbations. Neural Computation, 14(11), 2002.
[6] H. Jaeger and H. Haas. Harnessing nonlinearity: Predicting chaotic systems and saving energy
in wireless communication. Science, 304(5667):78?80, 2004.
[7] S. Bornholdt and T. R?ohl. Self-organized critical neural networks. Physical Review E, 67:066118,
2003.
[8] N. Bertschinger and T. Natschl?ager.
Supplementary information to the meanfield theory for randomly connected recurrent networks of threshold gates, 2004.
http://www.igi.tugraz.at/tnatschl/edge-of-chaos/mean-field-supplement.pdf.
| 2671 |@word bf:2 open:1 simulation:1 crucially:1 thereby:1 solid:3 initial:9 series:10 contains:2 liquid:3 activation:2 must:2 subsequent:1 visible:1 numerical:3 analytic:1 leipzig:1 plot:5 designed:1 update:4 discovering:1 accordingly:3 provides:1 math:1 node:12 lsm:1 ohl:1 consists:1 combine:1 expected:2 behavior:2 mpg:1 nor:1 p1:1 themselves:1 integrator:1 terminal:1 inspired:1 actual:2 increasing:2 provided:1 estimating:1 underlying:1 project:1 panel:4 circuit:1 developed:5 temporal:2 esti:4 computa:1 legi:1 exactly:1 classifier:4 uk:1 planck:1 before:1 local:2 tends:1 encoding:1 analyzing:1 firing:2 interpolation:1 black:3 emphasis:1 therein:2 suggests:1 range:1 averaged:8 decided:1 implement:1 differs:1 chaotic:24 projection:1 suggest:1 close:3 put:1 influence:1 impossible:1 applying:2 www:1 equivalent:1 map:6 center:1 yt:3 annealed:1 starting:1 independently:1 automaton:1 focused:1 react:2 rule:16 notion:1 autonomous:3 target:3 play:1 controlling:1 enhanced:1 hypothesis:3 asymmetric:1 predicts:1 observed:1 role:1 enters:1 calculate:2 graz:3 readout:7 connected:5 decrease:1 highest:3 rq:1 mentioned:1 principled:1 complexity:6 dynamic:25 signature:1 trained:2 depend:2 crit:1 serve:1 easily:1 differently:1 various:1 train:1 describe:2 europhys:1 harnessing:1 whose:1 quite:1 supplementary:1 say:1 otherwise:1 statistic:5 itself:1 online:1 propose:3 adaptation:1 tu:2 relevant:1 gen:1 achieve:1 amplified:1 requirement:1 jaeger:2 produce:1 sity:1 depending:1 recurrent:8 ac:1 derrida:4 develop:1 measured:4 minor:1 soc:10 implemented:1 implies:1 thick:1 closely:1 correct:1 filter:2 successor:1 enable:1 require:1 suffices:1 biological:2 adjusted:1 extension:1 physica:1 sufficiently:1 considered:6 mapping:1 vary:1 early:1 purpose:4 weighted:1 clearly:3 gaussian:3 always:1 rather:1 natschlaeger:1 pn:3 varying:1 timevarying:1 encode:1 focus:1 properly:1 consistently:1 prevalent:1 check:1 contrast:5 sense:2 glass:1 tional:1 unlikely:1 spurious:1 relation:2 ical:1 wij:9 transformed:1 germany:1 overall:2 flexible:1 pascal:1 denoted:1 animal:1 mutual:2 apriori:1 field:7 equal:2 never:1 having:1 saving:1 flipped:2 broad:1 future:1 others:1 randomly:10 composed:1 simultaneously:1 delayed:2 replaced:1 phase:16 consisting:1 fire:1 n1:1 organization:1 investigate:1 evaluation:1 adjust:5 analyzed:1 activated:1 accurate:2 edge:11 fu:2 necessary:2 ager:4 initialized:3 desired:1 theoretical:1 increased:1 boolean:3 deviation:1 predictor:3 delay:1 successful:1 conducted:1 reported:1 considerably:1 density:1 peak:2 stay:1 physic:1 off:1 together:1 continuously:1 connectivity:1 again:1 nm:3 langton:1 external:5 account:1 de:1 sec:9 summarized:1 caused:2 igi:2 depends:2 stream:1 onset:1 later:1 performed:2 tion:1 analyze:4 reached:1 start:2 capability:13 parallel:2 slope:2 bornholdt:1 spin:1 variance:2 spaced:1 conceptually:1 mc:4 confirmed:3 drive:2 history:1 phys:1 whenever:1 synaptic:7 definition:1 energy:1 mi:6 pbf:14 hamming:8 adjusting:1 austria:2 ut:21 improves:1 organized:6 sophisticated:1 actually:1 higher:2 dt:14 supervised:1 evaluated:1 furthermore:3 correlation:1 hand:2 receives:1 su:3 nonlinear:4 hopefully:1 indicated:4 gray:2 grows:1 facilitate:1 normalized:2 adequately:1 analytically:1 evolution:6 hence:6 symmetric:1 nonzero:2 maass:2 white:2 self:7 noted:1 steady:4 die:1 ut0:1 pdf:1 demonstrate:1 image:2 chaos:16 novel:1 recently:1 consideration:1 physical:2 exponentially:1 discussed:1 slight:1 refer:1 composition:1 mathematics:1 nonlinearity:1 had:1 robot:1 access:2 stable:2 deduce:2 dominant:1 recent:2 driven:5 certain:1 binary:1 vt:1 accomplished:1 seen:1 determine:4 signal:10 dashed:1 rv:1 calculation:1 long:2 pomeau:2 equally:1 coded:2 iteration:3 kernel:1 limt:3 achieved:3 background:2 whereas:1 want:1 interval:1 grow:1 rest:1 natschl:4 induced:1 tend:1 call:1 near:6 xj:3 timesteps:6 idea:4 regarding:1 t0:1 whether:1 speaking:1 useful:1 detailed:1 clear:1 processed:1 http:1 dotted:1 sign:1 estimated:2 ist:1 reformulation:1 threshold:6 drawn:5 changing:1 neither:1 destroy:1 relaxation:1 run:2 injected:1 respond:1 arrive:1 almost:1 separation:19 legenstein:1 dy:1 scaling:5 bit:14 fold:1 activity:11 fading:5 software:1 encodes:1 aspect:1 argument:1 separable:1 describes:2 evolves:1 biologically:1 intuitively:2 restricted:1 pr:2 den:1 taken:1 remains:1 eventually:1 needed:1 know:1 flip:2 serf:1 available:1 appropriate:1 generic:1 gate:4 thomas:2 assumes:3 top:1 denotes:4 running:4 tugraz:1 somewhere:1 especially:1 already:1 quantity:1 spike:1 occurs:2 dependence:3 exhibit:2 distance:15 separate:2 mapped:1 capacity:3 w0:2 haas:1 relationship:2 setup:1 robert:1 potentially:1 perform:2 diamond:1 upper:2 neuron:4 finite:1 timeseries:1 immediate:1 criticality:7 extended:1 looking:1 communication:1 perturbation:1 sharp:1 arbitrary:1 competence:1 community:1 pair:2 extensive:2 connection:1 optimized:1 able:5 suggested:1 bar:1 dynamical:4 below:2 regime:10 rf:1 max:1 memory:9 explanation:2 power:6 critical:29 suitable:1 meanfield:1 predicting:1 numerous:1 started:1 carried:1 mediated:1 extract:1 func:1 review:1 acknowledgement:1 fully:2 versus:1 integrate:1 degree:1 sufficient:1 article:1 principle:2 row:4 summary:2 supported:1 parity:8 last:2 wireless:1 allow:2 institute:2 markram:1 leaky:1 calculated:2 xn:1 transition:9 cumulative:1 rich:1 evaluating:1 sensory:1 forward:1 lett:1 adaptive:1 employing:1 keep:1 active:1 incoming:2 equ:2 xi:10 search:1 additionally:2 composing:1 complex:6 european:1 domain:5 allowed:2 x1:1 fig:12 position:1 vanish:1 formula:2 xt:9 specific:3 evidence:2 exists:1 importance:1 supplement:1 bertschinger:3 margin:2 suited:1 led:1 logarithmic:1 ordered:22 applies:1 marked:2 towards:6 fut:1 change:4 hard:1 determined:4 typical:1 specifically:1 nil:1 called:5 tnatschl:1 formally:1 internal:3 mark:1 support:2 assessed:1 tested:1 scratch:1 |
1,841 | 2,672 | Exponential Family Harmoniums
with an Application to Information Retrieval
Max Welling & Michal Rosen-Zvi
Information and Computer Science
University of California
Irvine CA 92697-3425 USA
[email protected]
Geoffrey Hinton
Department of Computer Science
University of Toronto
Toronto, 290G M5S 3G4, Canada
[email protected]
Abstract
Directed graphical models with one layer of observed random variables
and one or more layers of hidden random variables have been the dominant modelling paradigm in many research fields. Although this approach has met with considerable success, the causal semantics of these
models can make it difficult to infer the posterior distribution over the
hidden variables. In this paper we propose an alternative two-layer model
based on exponential family distributions and the semantics of undirected models. Inference in these ?exponential family harmoniums? is
fast while learning is performed by minimizing contrastive divergence.
A member of this family is then studied as an alternative probabilistic
model for latent semantic indexing. In experiments it is shown that they
perform well on document retrieval tasks and provide an elegant solution
to searching with keywords.
1 Introduction
Graphical models have become the basic framework for generative approaches to probabilistic modelling. In particular models with latent variables have proven to be a powerful
way to capture hidden structure in the data. In this paper we study the important subclass
of models with one layer of observed units and one layer of hidden units.
Two-layer models can be subdivided into various categories depending on a number of
characteristics. An important property in that respect is given by the semantics of the
graphical model: either directed (Bayes net) or undirected (Markov random field). Most
two-layer models fall in the first category or are approximations derived from it: mixtures
of Gaussians (MoG), probabilistic PCA (pPCA), factor analysis (FA), independent components analysis (ICA), sigmoid belief networks (SBN), latent trait models, latent Dirichlet
allocation (LDA, otherwise known as multinomial PCA, or mPCA) [1], exponential family PCA (ePCA), probabilistic latent semantic indexing (pLSI) [6], non-negative matrix
factorization (NMF), and more recently the multiple multiplicative factor model (MMF)
[8].
Directed models enjoy important advantages such as easy (ancestral) sampling and easy
handling of unobserved attributes under certain conditions. Moreover, the semantics of
directed models dictates marginal independence of the latent variables, which is a suitable
modelling assumption for many datasets. However, it should also be noted that directed
models come with an important disadvantage: inference of the posterior distribution of the
latent variables given the observations (which is, for instance, needed within the context of
the EM algorithm) is typically intractable resulting in approximate or slow iterative procedures. For important applications, such as latent semantic indexing (LSI), this drawback
may have serious consequences since we would like to swiftly search for documents that
are similar in the latent topic space.
A type of two-layer model that has not enjoyed much attention is the undirected analogue
of the above described family of models. It was first introduced in [10] where it was named
?harmonium?. Later papers have studied the harmonium under various names (the ?combination machine? in [4] and the ?restricted Boltzmann machine? in [5]) and turned it into
a practical method by introducing efficient learning algorithms. Harmoniums have only
been considered in the context of discrete binary variables (in both hidden and observed
layers), and more recently in the Gaussian case [7]. The first contribution of this paper is to
extend harmoniums into the exponential family which will make them much more widely
applicable.
Harmoniums also enjoy a number of important advantages which are rather orthogonal to
the properties of directed models. Firstly, their product structure has the ability to produce
distributions with very sharp boundaries. Unlike mixture models, adding a new expert may
decrease or increase the variance of the distribution, which may be a major advantage in
high dimensions. Secondly, unlike directed models, inference in these models is very fast,
due to the fact that the latent variables are conditionally independent given the observations.
Thirdly, the latent variables of harmoniums produce distributed representations of the input.
This is much more efficient than the ?grandmother-cell? representation associated with
mixture models where each observation is generated by a single latent variable. Their most
important disadvantage is the presence of a global normalization factor which complicates
both the evaluation of probabilities of input vectors1 and learning free parameters from
examples. The second objective of this paper is to show that the introduction of contrastive
divergence has greatly improved the efficiency of learning and paved the way for large scale
applications.
Whether a directed two-layer model or a harmonium is more appropriate for a particular application is an interesting question that will depend on many factors such as prior
(conditional) independence assumptions and/or computational issues such as efficiency of
inference. To expose the fact that harmoniums can be viable alternatives to directed models we introduce an entirely new probabilistic extension of latent semantic analysis (LSI)
[3] and show its usefulness in various applications. We do not want to claim superiority
of harmoniums over their directed cousins, but rather that harmoniums enjoy rather different advantages that deserve more attention and that may one day be combined with the
advantages of directed models.
2 Extending Harmoniums into the Exponential Family
Let xi , i = 1...Mx be the set of observed random variables and hj , j = 1...Mh be the set
of hidden (latent) variables. Both x and h can take values in either the continuous or the
discrete domain. In the latter case, each variable has states a = 1...D.
To construct an exponential family harmonium (EFH) we first choose Mx independent
distributions pi (xi ) for the observed variables and Mh independent distributions pj (hj )
1
However, it is easy to compute these probabilities up to a constant so it is possible to compare
probabilities of data-points.
for the hidden variables from the exponential family and combine them multiplicatively,
Mx
Y
p({xi }) =
ri (xi ) exp [
X
Mh
Y
p({hj }) =
?ia fia (xi ) ? Ai ({?ia })]
(1)
a
i=1
sj (hj ) exp [
X
j=1
?jb gjb (hj ) ? Bj ({?jb })]
(2)
b
where {fia (xi ), gjb (hj )} are the sufficient statistics for the models (otherwise known as
features), {?ia , ?jb } the canonical parameters of the models and {Ai , Bj } the log-partition
functions (or log-normalization factors). In the following we will consider log(ri (xi )) and
log(sj (hj )) as additional features multiplied by a constant.
Next, we couple the random variables in the log-domain by the introduction of a quadratic
interaction term,
X
X
X jb
p({xi , hj }) ? exp [
?ia fia (xi ) +
?jb gjb (hj ) +
Wia fia (xi )gjb (hj )]
(3)
ia
jb
ijab
Note that we did not write the log-partition function for this joint model in order to indicate our inability to compute it in general. For some combinations of exponential family
jb
distributions it may be necessary to restrict the domain of Wia
in order to maintain norjb
jb
malizability of the joint probability distribution (e.g. Wia ? 0 or Wia
? 0). Although
we could also have mutually coupled the observed variables (and/or the hidden variables)
using similar interaction terms we refrain from doing so in order to keep the learning and
inference procedures efficient. Consequently, by this construction the conditional probability distributions are a product of independent distributions in the exponential family with
shifted parameters,
p({xi }|{hj }) =
p({hj }|{xi }) =
Mx
Y
exp [
X
??ia fia (xi ) ? Ai ({??ia })]
i=1
a
Mh
Y
X
exp [
j=1
??ia = ?ia +
X
jb
Wia
gjb (hj ) (4)
jb
? jb gjb (hj ) ? Bj ({?
? jb })] ?
? jb = ?jb +
?
X
jb
Wia
fia (xi )(5)
ia
b
P
P
Finally, using the following identity, y exp a ?a fa (y) = exp A({?a }) we can also
compute the marginal distributions of the observed and latent variables,
X
X
X jb
p({xi }) ? exp [
?ia fia (xi ) +
Bj ({?jb +
Wia fia (xi )})]
(6)
ia
p({hj }) ? exp [
X
jb
j
?jb gjb (hj ) +
X
i
ia
Ai ({?ia +
X
jb
Wia
gjb (hj )})]
(7)
jb
Note that 1) we can only compute the marginal distributions up to the normalization constant and 2) in accordance with the semantics of undirected models, there is no marginal
independence between the variables (but rather conditional independence).
2.1 Training EF-Harmoniums using Contrastive Divergence
Let p?({xi }) denote the data distribution (or the empirical distribution in case we observe a
finite dataset), and p the model distribution. Under the maximum likelihood objective the
learning rules for the EFH are conceptually simple2 ,
? jb )ip? ? hB 0 (?
? jb )ip
??ia ? hfia (xi )ip? ? hfia (xi )ip
??jb ? hB 0 (?
(8)
jb
2
jb
These learning rules are derived by taking derivatives of the log-likelihood objective using Eqn.6.
0 ?
0 ?
?Wijab ? hfia (xi )Bjb
(?jb )ip? ? hfia (xi )Bjb
(?jb )ip
(9)
0
? jb )/? ?
? jb with ?
? jb defined in Eqn.5. One should note
where we have defined Bjb
= ?Bj (?
that these learning rules are changing the parameters in an attempt to match the expected
sufficient statistics of the data distribution and the model distribution (while maximizing
entropy). Their simplicity is somewhat deceptive, however, since the averages h?ip are
intractable to compute analytically and Markov chain sampling or mean field calculations
are typically wheeled out to approximate them. Both have difficulties: mean field can only
represent one mode of the distribution and MCMC schemes are slow and suffer from high
variance in their estimates.
In the case of binary harmoniums (restricted BMs) it was shown in [5] that contrastive divergence has the potential to greatly improve on the efficiency and reduce the variance of
the estimates needed in the learning rules. The idea is that instead of running the Gibbs
sampler to its equilibrium distribution we initialize Gibbs samplers on each data-vector
and run them for only one (or a few) steps in parallel. Averages h?ip in the learning rules
Eqns.8,9 are now replaced by averages h?ipCD where pCD is the distribution of samples
that resulted from the truncated Gibbs chains. This idea is readily generalized to EFHs.
Due to space limitations we refer to [5] for more details on contrastive divergence learning3 . Deterministic learning rules can also be derived straightforwardly by generalizing the
results described in [12] to the exponential family.
3 A Harmonium Model for Latent Semantic Indexing
To illustrate the new possibilities that have opened up by extending harmoniums to the
exponential family we will next describe a novel model for latent semantic indexing (LSI).
This will represent the undirected counterpart of pLSI [6] and LDA [1].
One of the major drawbacks of LSI is that inherently discrete data (word counts) are being
modelled with variables in the continuous domain. The power of LSI on the other hand is
that it provides an efficient mapping of the input data into a lower dimensional (continuous)
latent space that has the effect of de-noising the input data and inferring semantic relationships among words. To stay faithful to this idea and to construct a probabilistic model on
the correct (discrete) domain we propose the following EFH with continuous latent topic
variables, hj , and discrete word-count variables, xia ,
p({hj }|{xia }) =
Mh
Y
j=1
p({xia }|{hj }) =
Mx
Y
i=1
X j
Nhj [
Wia xia , 1]
(10)
ia
S{xia } [?ia +
X
j
hj Wia
]
(11)
j
P
Note that {xia } represent indicator variables satisfying a xia = 1 ?i, where xia = 1
means that word ?i? in the vocabulary was observed ?a? times. Nh [?, ?] denotes a normal
PD
distribution with mean ? and std.? and S{xa } [?a ] ? exp ( a=1 ?a xa ) is the softmax
function defining a probability distribution over x. Using Eqn.6 we can easily deduce the
marginal distribution of the input variables,
X
1X X j
(
Wia xia )2 ]
(12)
p({xia }) ? exp [
?ia xia +
2
j
ia
ia
3
Non-believers in contrastive divergence are invited to simply run the the Gibbs sampler to equilibrium before they do an update of the parameters. They will find that due to the special bipartite
structure of EFHs learning is still more efficient than for general Boltzmann machines.
j
We observe that the role of the components Wia
is that of templates or prototypes: input
P
j
vectors xia with large inner products ia Wia xia ?j will have high probability under this
model. Just like pLSI and LDA can be considered as natural generalizations of factor
analysis (which underlies LSI) into the class of directed models on the discrete domain, the
above model can be considered as the natural generalization of factor analysis into class
of undirected models on the discrete domain. This idea is supported by the result that the
same model with Gaussian units in both hidden and observed layers is in fact equivalent to
factor analysis [7].
3.1 Identifiability
From the form of the marginal distribution Eqn.12 we can derive a number of transformations of the parameters that will leave the distribution invariant. First we note that the comP
j
j
k
ponents Wia
can be rotated and mirrored arbitrarily in latent space4 : Wia
? k U jk Wia
T
with U U = I. Secondly, we note that observed variables xia satisfy a constraint,
P
j
a xia = 1 ?i. This results in a combined shift invariance for the components Wia and
the offsets ?ia . Taken together, this results in the following set of transformations,
X
XX j
j
j
k
Wia
?
U jk (Wia
+ Vik )
?ia ? (?ia + ?i ) ?
(
Vl )(Wia
)
(13)
j
k
l
where U T U = I. Although these transformations leave the marginal distribution over the
observable variables invariant, they do change the latent representation and as such may
have an impact on retrieval performance (if we use a fixed similarity measure between
topic representations of documents). To fix the spurious degrees of freedom we have choP
j n
xia . First,
sen to impose conditions on the representations in latent space: hnj = ia Wia
we center the latent representations which has the effect of minimizing the ?activity? of
the latent variables and moving as much log-probability as possible to the constant component ?ia . Next we align the axes in latent space with the eigen-directions of the latent
covariance matrix. This has the effect of approximately decorrelating the marginal latent
activities. This follows because the marginal distribution in latent space can be approxP
P Q
j n
xia , 1]/N where we have used Eqn.10 and
imated by: p({hj }) ? n j Nhj [ ia Wia
replaced p({xia }) by its empirical distribution. Denoting by ? and ? = U T ?U the sample
mean and sample covariance of {hnj }, it is not hard to show that the following transformation will have the desired effect5 :
?
?
X
X
1 k
j
j
jk
k
?
?ia ? ?ia +
?j Wia
(14)
Wia ?
U
Wia ?
Mx
j
k
One could go one step further than the de-correlation process described above by introducing covariances ? in the conditional Gaussian distribution of the latent variables Eqn.10.
This would not result in a more general model because the effect of this on the marginal
P
j
k
KK T = ?.
distribution over the observed variables is given by: Wia
? k K jk Wia
However, the extra freedom can be used to define axes in latent space for which the projected data become approximately independent and have the same scale in all directions.
4
Technically we call this the Euclidean group of transformations.
P
Some spurious degrees of freedom remain since shifts ?i and shifts Vij that satisfy i Vij = 0
will not affect the projection into latent space. One could decide to fix the remaining degrees of
freedom by for
(subject
to
P in Lj2 norm
Pexample requiring that components are asjsmall asj possible
P
j
1
1
W
+
the constraint i Vij = 0), leading to the further shifts, Wia
? Wia ? D
W
ia
ia
a
ia
DMx
P
1
and ?ia ? ?ia ? D a ?ia .
5
Precision vs. Recall for Newsgroups Dataset
0.8
Precision vs. Recall for Newsgroups Dataset
Precision vs. Recall for Newgroups Dataset
0.8
0.65
0.6
EFH
0.7
EFH+MF
LSA
EFH
0.7
EFH+MF
0.55
LSA
0.6
0.6
EFH
0.4
TFIDF
Precision
Precision
Precision
0.5
0.45
0.5
0.35
TFIDF
0.2
0.1 ?5
10
?4
10
0.3
0.25
RANDOM
0.2
?3
10
?2
10
Recall
(a)
?1
10
0
10
0.15 ?5
10
TFIDF
0.4
0.3
0.3
LSA
0.5
0.4
0.2
RANDOM
?4
10
?3
10
Recall
(b)
?2
10
?1
10
0
10
0.1 ?5
10
RANDOM
?4
10
?3
10
?2
10
?1
10
0
10
Recall
(c)
Figure 1: Precision-recall curves when the query was (a) entire documents, (b) 1 keyword, (c) 2
keywords for the EFH with and without 10 MF iterations, LSI , TFIDF weighted words and random
guessing. PR curves with more keywords looked very similar to (c). A marker at position k (counted
from the left along a curve) indicates that 2k?1 documents were retrieved.
4 Experiments
Newsgroups:
We have used the reduced version of the ?20newsgroups? dataset prepared for MATLAB
by Roweis6 . Documents are presented as 100 dimensional binary occurrence vectors and
tagged as a member of 1 out of 4 domains. Documents contain approximately 4% of the
words, averaged across the 16242 postings.
An EFH model with 10 latent variables was trained on 12000 training cases using stochastic gradient descent on mini-batches of 1000 randomly chosen documents (training
time approximately 1 hour on a 2GHz PC). A momentum term was added to speed up
convergence. To test the quality of the trained model we mapped the remaining 4242
P
j
j
xia and where {Wia
, ?ia } were
query documents into latent space using hj = ia Wia
?gauged? as in Eqns.14. Precision-recall curves were computed by comparing training
and query documents using the usual ?cosine coefficient? (cosine of the angle between
documents) and reporting success when the retrieved document was in the same domain
as the query (results averaged over all queries). In figure 1a we compare the results
with LSI (also 10 dimensions) [3] where we preprocessed the data in the standard way
(x ? log(1 + x) and entropy weighting of the words) and to similarity in word space
using TF-IDF weighting of the words. In figure 1b,c we show PR curves when only 1 or
2 keywords were provided corresponding to randomly observed words in the query document. The EFH model allows a principled way to deal with unobserved entries by inferring
them using the model (in all other methods we insert 0 for the unobserved entries which
corresponds to ignoring
them). We have used a few iterations of mean field to achieve
P P
k
k
Wjb
+ ?jb )?
xjb ]/?i where ?i is a normalization constant
that: x
?ia ? exp [ jb ( k Wia
PD
?ia = 1 ?i. We note that this is
and where x
?ia represent probabilities: x
?ia ? [0, 1],
a=1 x
still highly efficient and achieves a significant improvement in performance. In all cases we
find that without any preprocessing or weighting EFH still outperforms the other methods
except when large numbers of documents were retrieved.
In the next experiment we compared performance of EFH, LSI and LDA by training models on a random subset of 15430 documents with 5 and 10 latent dimensions (this was
found to be close to optimal for LDA). The EFH and LSI models were trained as in the
previous experiment while the training and testing details7 for LDA can be found in [9].
For the remaining test documents we clamped a varying number of observed words and
6
7
http://www.cs.toronto.edu/?roweis/data.html
The approximate inference procedure was implemented using Gibbs sampling.
Latent Representations of Newsgroups Data
8
6
0.25
EFH+MF 10D
EFH+MF 5D
4
?4
0.2
LSI 5D
0.15
LDA 5D
LDA 10D
?2
0
2
4
0.1
LSI 10D
6
8
0.05
10
0
0
2
4
6
number of keywords
(a)
8
10
12
6
4
2
0
?2
?4
?6
(b)
Retrieval Performance for NIPS Dataset
fraction documents also retrieved by TF?IDF
fraction observed words correctly predicted
Document Reconstruction for Newsgroup Dataset
0.3
0.52
0.5
0.48
0.46
0.44
0.42
0.4
0.38
0
20
40
60
80
number of retrieved documents
100
(c)
Figure 2: (a) Fraction of observed words that was correctly observed by EFH, LSI and LDA using 5
and 10 latent variables when we vary the number of keywords (observed words that were ?clamped?),
(b) latent 3-D representations of newsgroups data, (c) Fraction of documents retrieved by EFH on the
NIPS dataset which was also retrieved by the TF-IDF method.
asked the models to predict the remaining observed words in the documents by computing the probabilities for all words in the vocabulary to be present and ranking them (see
previous paragraph for details). By comparing the list of the R remaining observed words
in the document with the top-R ranked inferred words we computed the fraction of correctly predicted words. The results are shown in figure 2a as a function of the number of
clamped words. To provide anecdotal evidence that EFH can infer semantic relationships
we clamped the words ?drive? ?driver? and ?car? which resulted in: ?car? ?drive? ?engine?
?dealer? ?honda? ?bmw? ?driver? ?oil? as the most probable words in the documents. Also,
clamping ?pc? ?driver? and ?program? resulted in: ?windows? ?card? ?dos? ?graphics? ?software? ?pc? ?program? ?files?.
NIPS Conference Papers:
Next we trained a model with 5 latent dimensions on the NIPS dataset8 which has a large
vocabulary size (13649 words) and contains 1740 documents of which 1557 were used
for training and 183 for testing. Count values were redistributed in 12 bins. The array
W contains therefore 5 ? 13649 ? 12 = 818940 parameters. Training was completed in the
order of a few days. Due to the lack of document labels it is hard to assess the quality of the
trained model. We choose to compare performance on document retrieval with the ?golden
standard?: cosine similarity in TF-IDF weighted word space. In figure 2c we depict the
fraction of documents retrieved by EFH that was also retrieved by TF-IDF as we vary the
number of retrieved documents. This correlation is indeed very high but note that EFH
computes similarity in a 5-D space while TF-IDF computes similarity in a 13649-D space.
5 Discussion
The main point of this paper was to show that there is a flexible family of 2-layer probabilistic models that represents a viable alternative to 2-layer causal (directed) models. These
models enjoy very different properties and can be trained efficiently using contrastive divergence. As an example we have studied an EFH alternative for latent semantic indexing
where we have found that the EFH has a number of favorable properties: fast inference allowing fast document retrieval and a principled approach to retrieval with keywords. These
were preliminary investigations and it is likely that domain specific adjustments such as a
more intelligent choice of features or parameterization could further improve performance.
Previous examples of EFH include the original harmonium [10], Gaussian variants thereof
[7], and the PoT model [13] which couples a gamma distribution with the covariance of a
8
Obtained from http://www.cs.toronto.edu/?roweis/data.html.
normal distribution. Some exponential family extensions of general Boltzmann machines
were proposed in [2], [14], but they do not have the bipartite structure that we study here.
While the components of the Gaussian-multinomial EFH act as prototypes or templates
for highly probable input vectors, the components of the PoT act as constraints (i.e. input
vectors with large inner product have low probability). This can be traced back to the
shape of the non-linearity B in Eqn.6. Although by construction B must be convex (it
is the log-partition function), for large input values it can both be positive (prototypes,
e.g. B(x) = x2 ) or negative (constraints, e.g. B(x) = ?log(1 + x)). It has proven
difficult to jointly model both prototypes and constraints in the this formalism except for
the fully Gaussian case [11]. A future challenge is therefore to start the modelling process
with the desired non-linearity and to subsequently introduce auxiliary variables to facilitate
inference and learning.
References
[1] D. M. Blei, A. Y. Ng, and M. I. Jordan. Latent Dirichlet allocation. Journal of Machine Learning
Research, 3:993?1022, 2003.
[2] C.K.I.Williams. Continuous valued Boltzmann machines. Technical report, 1993.
[3] S.C. Deerwester, S.T. Dumais, T.K. Landauer, G.W. Furnas, and R.A. Harshman. Indexing by
latent semantic analysis. Journal of the American Society of Information Science, 41(6):391?
407, 1990.
[4] Y. Freund and D. Haussler. Unsupervised learning of distributions of binary vectors using
2-layer networks. In Advances in Neural Information Processing Systems, volume 4, pages
912?919, 1992.
[5] G.E. Hinton. Training products of experts by minimizing contrastive divergence. Neural Computation, 14:1771?1800, 2002.
[6] Thomas Hofmann. Probabilistic latent semantic analysis. In Proc. of Uncertainty in Artificial
Intelligence, UAI?99, Stockholm, 1999.
[7] T. K. Marks and J. R. Movellan. Diffusion networks, products of experts, and factor analysis.
Technical Report UCSD MPLab TR 2001.02, University of California San Diego, 2001.
[8] B. Marlin and R. Zemel. The multiple multiplicative factor model for collaborative filtering. In
Proceedings of the 21st International Conference on Machine Learning, volume 21, 2004.
[9] M. Rosen-Zvi, T. Griffiths, M. Steyvers, and P. Smyth. The author-topic model for authors
and documents. In Proceedings of the Conference on Uncertainty in Artificial Intelligence,
volume 20, 2004.
[10] P. Smolensky. Information processing in dynamical systems: foundations of harmony theory.
In D.E. Rumehart and J.L. McClelland, editors, Parallel Distributed Processing: Explorations
in the Microstructure of Cognition. Volume 1: Foundations. McGraw-Hill, New York, 1986.
[11] M. Welling, F. Agakov, and C.K.I. Williams. Extreme components analysis. In Advances in
Neural Information Processing Systems, volume 16, Vancouver, Canada, 2003.
[12] M. Welling and G.E. Hinton. A new learning algorithm for mean field Boltzmann machines.
In Proceedings of the International Conference on Artificial Neural Networks, Madrid, Spain,
2001.
[13] M. Welling, G.E. Hinton, and S. Osindero. Learning sparse topographic representations with
products of student-t distributions. In Advances in Neural Information Processing Systems,
volume 15, Vancouver, Canada, 2002.
[14] R. Zemel, C. Williams, and M. Mozer. Lending direction to neural networks. Neural Networks,
8(4):503?512, 1995.
| 2672 |@word version:1 norm:1 efh:25 dealer:1 covariance:4 contrastive:8 gjb:8 tr:1 contains:2 denoting:1 document:30 outperforms:1 comparing:2 michal:1 must:1 readily:1 partition:3 hofmann:1 shape:1 update:1 depict:1 v:3 generative:1 intelligence:2 parameterization:1 blei:1 provides:1 lending:1 toronto:5 honda:1 firstly:1 along:1 become:2 driver:3 viable:2 combine:1 paragraph:1 lj2:1 introduce:2 g4:1 expected:1 indeed:1 ica:1 mplab:1 window:1 provided:1 xx:1 moreover:1 linearity:2 spain:1 unobserved:3 transformation:5 marlin:1 subclass:1 golden:1 act:2 unit:3 enjoy:4 superiority:1 lsa:3 harshman:1 before:1 positive:1 accordance:1 consequence:1 approximately:4 studied:3 deceptive:1 factorization:1 averaged:2 directed:13 practical:1 faithful:1 testing:2 movellan:1 procedure:3 empirical:2 dictate:1 projection:1 word:24 griffith:1 close:1 noising:1 context:2 www:2 equivalent:1 deterministic:1 center:1 maximizing:1 go:1 attention:2 williams:3 convex:1 simplicity:1 rule:6 haussler:1 array:1 steyvers:1 searching:1 construction:2 diego:1 smyth:1 satisfying:1 jk:4 std:1 agakov:1 observed:19 role:1 capture:1 pexample:1 sbn:1 keyword:1 decrease:1 principled:2 mozer:1 pd:2 asked:1 trained:6 depend:1 harmonium:19 technically:1 bipartite:2 efficiency:3 easily:1 mh:5 joint:2 various:3 fast:4 describe:1 query:6 artificial:3 paved:1 zemel:2 widely:1 valued:1 otherwise:2 ability:1 statistic:2 topographic:1 jointly:1 ip:8 advantage:5 net:1 sen:1 propose:2 reconstruction:1 interaction:2 product:7 redistributed:1 uci:1 turned:1 achieve:1 roweis:2 convergence:1 extending:2 produce:2 leave:2 rotated:1 depending:1 illustrate:1 derive:1 keywords:7 implemented:1 c:3 predicted:2 come:1 indicate:1 met:1 auxiliary:1 direction:3 pot:2 drawback:2 correct:1 attribute:1 opened:1 stochastic:1 subsequently:1 exploration:1 bin:1 subdivided:1 fix:2 generalization:2 microstructure:1 preliminary:1 investigation:1 tfidf:4 probable:2 stockholm:1 secondly:2 extension:2 insert:1 considered:3 ic:1 normal:2 exp:12 wheeled:1 equilibrium:2 mapping:1 bj:5 predict:1 claim:1 cognition:1 major:2 achieves:1 vary:2 favorable:1 proc:1 applicable:1 harmony:1 label:1 expose:1 tf:6 weighted:2 anecdotal:1 gaussian:6 rather:4 hj:23 varying:1 derived:3 ax:2 improvement:1 modelling:4 likelihood:2 indicates:1 greatly:2 inference:8 vl:1 typically:2 entire:1 dataset8:1 hidden:9 spurious:2 semantics:5 simple2:1 issue:1 among:1 html:2 flexible:1 softmax:1 initialize:1 special:1 marginal:10 field:6 construct:2 ng:1 sampling:3 represents:1 unsupervised:1 future:1 rosen:2 jb:33 report:2 intelligent:1 serious:1 few:3 randomly:2 gamma:1 divergence:8 resulted:3 replaced:2 maintain:1 attempt:1 freedom:4 imated:1 possibility:1 highly:2 evaluation:1 mixture:3 extreme:1 pc:3 chain:2 necessary:1 orthogonal:1 euclidean:1 desired:2 causal:2 complicates:1 instance:1 newgroups:1 formalism:1 disadvantage:2 introducing:2 entry:2 subset:1 usefulness:1 osindero:1 graphic:1 zvi:2 nhj:2 straightforwardly:1 swiftly:1 combined:2 dumais:1 st:1 international:2 ancestral:1 stay:1 probabilistic:8 together:1 choose:2 expert:3 derivative:1 leading:1 american:1 hnj:2 potential:1 de:2 student:1 coefficient:1 satisfy:2 ranking:1 performed:1 multiplicative:2 later:1 doing:1 start:1 bayes:1 parallel:2 identifiability:1 contribution:1 ass:1 collaborative:1 variance:3 characteristic:1 efficiently:1 conceptually:1 dmx:1 modelled:1 comp:1 drive:2 m5s:1 thereof:1 associated:1 couple:2 irvine:1 ppca:1 dataset:8 wjb:1 recall:8 car:2 back:1 day:2 improved:1 decorrelating:1 xa:2 just:1 correlation:2 hand:1 eqn:7 marker:1 lack:1 mode:1 lda:9 quality:2 oil:1 name:1 usa:1 effect:4 requiring:1 contain:1 counterpart:1 facilitate:1 analytically:1 tagged:1 semantic:11 deal:1 conditionally:1 eqns:2 chop:1 noted:1 cosine:3 generalized:1 hill:1 ef:1 recently:2 novel:1 sigmoid:1 multinomial:2 nh:1 thirdly:1 extend:1 volume:6 trait:1 refer:1 significant:1 gibbs:5 ai:4 enjoyed:1 moving:1 vectors1:1 similarity:5 deduce:1 align:1 dominant:1 posterior:2 plsi:3 retrieved:10 certain:1 binary:4 success:2 refrain:1 arbitrarily:1 additional:1 somewhat:1 impose:1 paradigm:1 multiple:2 infer:2 technical:2 match:1 calculation:1 retrieval:7 impact:1 underlies:1 basic:1 variant:1 mog:1 iteration:2 normalization:4 represent:4 cell:1 want:1 invited:1 extra:1 unlike:2 file:1 subject:1 elegant:1 undirected:6 member:2 jordan:1 call:1 presence:1 easy:3 hb:2 newsgroups:6 independence:4 affect:1 mmf:1 restrict:1 reduce:1 idea:4 prototype:4 inner:2 space4:1 vik:1 cousin:1 shift:4 whether:1 pca:3 suffer:1 york:1 matlab:1 prepared:1 category:2 mcclelland:1 reduced:1 http:2 lsi:13 canonical:1 mirrored:1 shifted:1 correctly:3 discrete:7 write:1 group:1 traced:1 epca:1 changing:1 preprocessed:1 pj:1 diffusion:1 fraction:6 deerwester:1 run:2 angle:1 powerful:1 uncertainty:2 named:1 reporting:1 family:16 decide:1 entirely:1 layer:14 quadratic:1 activity:2 constraint:5 pcd:1 idf:6 ri:2 software:1 x2:1 speed:1 department:1 combination:2 remain:1 across:1 em:1 restricted:2 indexing:7 invariant:2 pr:2 taken:1 mutually:1 count:3 needed:2 fia:8 gaussians:1 multiplied:1 observe:2 appropriate:1 occurrence:1 alternative:5 batch:1 eigen:1 original:1 thomas:1 denotes:1 dirichlet:2 running:1 remaining:5 top:1 graphical:3 completed:1 include:1 society:1 objective:3 question:1 added:1 looked:1 fa:2 usual:1 guessing:1 gradient:1 mx:6 mapped:1 card:1 topic:4 relationship:2 multiplicatively:1 kk:1 minimizing:3 mini:1 difficult:2 negative:2 boltzmann:5 perform:1 allowing:1 observation:3 markov:2 datasets:1 finite:1 descent:1 truncated:1 defining:1 hinton:5 ucsd:1 sharp:1 canada:3 nmf:1 inferred:1 introduced:1 california:2 engine:1 hour:1 nip:4 deserve:1 dynamical:1 smolensky:1 grandmother:1 challenge:1 program:2 max:1 belief:1 analogue:1 ia:41 suitable:1 power:1 difficulty:1 natural:2 ranked:1 indicator:1 scheme:1 improve:2 coupled:1 prior:1 vancouver:2 freund:1 fully:1 gauged:1 interesting:1 limitation:1 allocation:2 filtering:1 proven:2 geoffrey:1 foundation:2 degree:3 mpca:1 sufficient:2 editor:1 vij:3 pi:1 supported:1 free:1 fall:1 template:2 taking:1 sparse:1 distributed:2 ghz:1 boundary:1 dimension:4 xia:19 vocabulary:3 curve:5 computes:2 author:2 projected:1 preprocessing:1 bm:1 counted:1 san:1 welling:5 sj:2 approximate:3 observable:1 asj:1 mcgraw:1 keep:1 global:1 uai:1 xi:22 landauer:1 search:1 latent:42 iterative:1 continuous:5 wia:32 ca:1 inherently:1 ignoring:1 domain:10 did:1 main:1 madrid:1 slow:2 furnas:1 precision:8 inferring:2 position:1 momentum:1 exponential:13 clamped:4 weighting:3 posting:1 specific:1 offset:1 list:1 evidence:1 intractable:2 adding:1 clamping:1 mf:5 entropy:2 generalizing:1 simply:1 likely:1 adjustment:1 corresponds:1 conditional:4 identity:1 bmw:1 consequently:1 ponents:1 considerable:1 change:1 hard:2 except:2 sampler:3 invariance:1 newsgroup:1 mark:1 latter:1 inability:1 mcmc:1 handling:1 |
1,842 | 2,673 | Convergence and No-Regret in Multiagent
Learning
Michael Bowling
Department of Computing Science
University of Alberta
Edmonton, Alberta
Canada T6G 2E8
[email protected]
Abstract
Learning in a multiagent system is a challenging problem due to two key
factors. First, if other agents are simultaneously learning then the environment is no longer stationary, thus undermining convergence guarantees. Second, learning is often susceptible to deception, where the other
agents may be able to exploit a learner?s particular dynamics. In the
worst case, this could result in poorer performance than if the agent was
not learning at all. These challenges are identifiable in the two most common evaluation criteria for multiagent learning algorithms: convergence
and regret. Algorithms focusing on convergence or regret in isolation
are numerous. In this paper, we seek to address both criteria in a single
algorithm by introducing GIGA-WoLF, a learning algorithm for normalform games. We prove the algorithm guarantees at most zero average
regret, while demonstrating the algorithm converges in many situations
of self-play. We prove convergence in a limited setting and give empirical results in a wider variety of situations. These results also suggest
a third new learning criterion combining convergence and regret, which
we call negative non-convergence regret (NNR).
1
Introduction
Learning to select actions to achieve goals in a multiagent setting requires overcoming a
number of key challenges. One of these challenges is the loss of the stationarity assumption
when multiple agents are learning simultaneously. Another challenge is guaranteeing that
the learner cannot be deceptively exploited by another agent. Both of these challenges distinguish the multiagent learning problem from traditional single-agent learning, and have
been gaining recent attention as multiagent applications continue to proliferate.
In single-agent learning tasks, it is reasonable to assume that the same action from the
same state will result in the same distribution over outcomes, both rewards and next states.
In other words, the environment is stationary. In a multiagent task with other learning
agents, the outcomes of an agent?s action will vary with the changing policies of the other
agents. Since most of the convergence results in reinforcement learning depend upon the
environment being stationary, convergence is often difficult to obtain in multiagent settings.
The desirability of convergence has been recently contested. We offer some brief insight
into this debate in the introduction of the extended version of this paper [1].
Equilibrium learners [2, 3, 4] are one method of handling the loss of stationarity. These algorithms learn joint-action values, which are stationary, and in certain circumstances guarantee these values converge to Nash (or correlated) equilibrium values. Using these values,
the player?s strategy corresponds to the player?s component of some Nash (or correlated)
equilibrium. This convergence of strategies is guaranteed nearly independently of the actions selected by the other agents, including when other agents play suboptimal responses.
Equilibrium learners, therefore, can fail to learn best-response policies even against simple
non-learning opponents.1 Best-response learners [5, 6, 7] are an alternative approach that
has sought to learn best-responses, but still considering whether the resulting algorithm
converges in some form. These approaches usually examine convergence in self-play, and
have included both theoretical and experimental results.
The second challenge is the avoidance of exploitation. Since learning strategies dynamically change their action selection over time, it is important to know that the change cannot
be exploited by a clever opponent. A deceptive strategy may ?lure? a dynamic strategy
away from a safe choice in order to switch to a strategy where the learner receives much
lower reward. For example, Chang and Kaelbling [8] demonstrated that the best-response
learner PHC [7] could be exploited by a particular dynamic strategy. One method of measuring whether an algorithm can be exploited is the notion of regret. Regret has been
explored both in game theory [9] and computer science [10, 11]. Regret measures how
much worse an algorithm performs compared to the best static strategy, with the goal to
guarantee at least zero average regret, no-regret, in the limit.
These two challenges result in two completely different criteria for evaluation: convergence and no-regret. In addition, they have almost exclusively been explored in isolation.
For example, equilibrium learners can have arbitrarily large average regret. On the other
hand, no-regret learners? strategies rarely converge in self-play [12] in even the simplest
of games.2 In this paper, we seek to explore these two criteria in a single algorithm for
learning in normal-form games. In Section 2 we present a more formal description of the
problem and the two criteria. We also examine key related work in applying gradient ascent
algorithms to this learning problem. In Section 3 we introduce GIGA-WoLF, an algorithm
with both regret and convergence properties. The algorithm is followed by theoretical and
experimental analyses in Sections 4 and 5, respectively, before concluding.
2
Online Learning in Games
A game in normal form is a tuple, (n, A1...n , R1...n ), where n is the number of players in
the game, Ai is a set of actions available to player i (A = A1 ?. . . ?An ), and Ri : A ? <
is a mapping from joint actions to player i?s reward. The problem of learning in a normalform game is one of repeatedly selecting an action and receiving a reward, with a goal of
maximizing average reward against an unknown opponent. If there are two players then it
is convenient to write a player?s reward function as a |A1 | ? |A2 | matrix. Three example
normal-form games are shown in Table 1.
Unless stated otherwise we will assume the learning algorithm is player one. In the context
of a particular learning algorithm and a particular opponent, let rt ? <|A1 | be the vector
of actual rewards that player one would receive at time t for each of its actions. Let x t ?
1
This work is not restricted to zero-sum games and our use of the word ?opponent? refers simply
to other players in the game.
2
A notable exception is Hart and Mas-Colell?s algorithm that guarantees the empirical distribution of play converges to that of a correlated equilibrium. Neither strategies nor expected values
necessarily converge, though.
Table 1: Examples of games in normal-form.
H
H
R1 =
T
?
AB
T
1 ?1
?1 1
?
R2 = ?R1
(a) Matching Pennies
R1 =
A
B
?
R2 =
A
B
?
R
0 3
1 2
?
3 2
0 1
?
(b) Tricky Game
P
S
0
1
R
0 ?1 1
R1 = P @ 1 0 ?1 A
S
?1 1 0
R2 = ?R1
(c) Rock?Paper?Scissors
P D(A1 ) be the algorithm?s strategy at time t, selected from probability distributions over
actions. So, player one?s expected payoff at time t is (rt ? xt ). Let 1a be the probability
distribution that assigns probability 1 to action a ? A1 . Lastly, we will assume the reward
2
for any action is bounded by rmax and therefore ||rt ||2 ? |A1 |rmax
.
2.1
Evaluation Criteria
One common evaluation criterion for learning in normal-form games is convergence. There
are a number of different forms of convergence that have been examined
P in the literature.
These include, roughly increasingPin strength: average reward (i.e., (rt ? xt )/T ), empirical distribution of actions (i.e.,
xt /T ), expected reward (i.e., (rt ? xt )), and strategies
(i.e., xt ). We focus in this paper on convergence of strategies as this implies the other three
forms of convergence as well. In particular, we will say an algorithm converges against a
particular opponent if and only if limt?? xt = x? .
The second common evaluation criterion is regret. Total regret3 is the difference between
the maximum total reward of any static strategy given the past history of play and the
algorithm?s total reward.
RT ? max
a?A1
T
X
t=1
((rt ? 1a ) ? (rt ? xt ))
Average regret is just the asymptotic average of total regret, limT ?? RT /T . An algorithm
has no-regret if and only if the average regret is less than or equal to zero against all opponent strategies. The no-regret property makes a strong claim about the performance of
the algorithm: the algorithm?s expected average reward is at least as large as the expected
average award any static strategy could have achieved. In other words, the algorithm is
performing at least as well as any static strategy.
2.2
Gradient Ascent Learning
Gradient ascent is a simple and common technique for finding parameters that optimize
a target function. In the case of learning in games, the parameters represent the player?s
strategy, and the target function is expected reward. We will examine three recent results
evaluating gradient ascent learning algorithms in normal-form games.
Singh, Kearns, and Mansour [6] analyzed gradient ascent (IGA) in two-player, two-action
games, e.g., Table 1(a) and (b). They examined the resulting strategy trajectories and payoffs in self-play, demonstrating that strategies do not always converge to a Nash equilibrium, depending on the game. They proved, instead, that average payoffs converge (a
3
Our analysis focuses on expectations of regret (total and average), similar to [10, 11]. Although
note that for any self-oblivious behavior, including GIGA-WoLF, average regret of at most zero on
expectation implies universal consistency, i.e., regret of at most with high probability [11].
weaker form of convergence) to the payoffs of the equilibrium. WoLF-IGA [7] extended
this work to the stronger form of convergence, namely convergence of strategies, through
the use of a variable learning rate. Using the WoLF (?Win or Learn Fast?) principle, the algorithm would choose a larger step size when the current strategy had less expected payoff
than the equilibrium strategy. This results in strategies converging to the Nash equilibrium
in a variety of games including all two-player, two-action games.4 Zinkevich [11] looked
at gradient ascent using the evaluation criterion of regret. He first extended IGA beyond
two-player, two-action games. His algorithm, GIGA (Generalized Infinitesimal Gradient
Ascent), updates strategies using an unconstrained gradient, and then projects the resulting
strategy vector back into the simplex of legal probability distributions,
xt+1 = P (xt + ?t rt )
where
P (x) = argmin ||x ? x0 ||,
(1)
x0 ?P D(A1 )
?t is the stepsize at time t, and || ? || is the standard L2 norm. He proved GIGA?s total regret
is bounded by,
?
?
2
RT ?
T + |A|rmax
( T ? 1/2).
(2)
Since GIGA is identical to IGA in two-player, two-action games, we also have that GIGA
achieves the weak form of convergence in this subclass of games. It is also true, though,
that GIGA?s strategies do not converge in self-play even in simple games like matching
pennies.
In the next section, we present an algorithm that simultaneously achieves GIGA?s no-regret
result and part of WoLF-IGA?s convergence result. We first present the algorithm and then
analyze these criteria both theoretically and experimentally.
3
GIGA-WoLF
GIGA-WoLF is a gradient based learning algorithm that internally keeps track of two different gradient-updated strategies, xt and zt . The algorithm chooses actions according to
the distribution xt , but updates both xt and zt after each iteration. The update rules consist
of three steps.
(1) x
?t+1
(2) zt+1
?t+1
(3) xt+1
= P (xt + ?t rt )
= P (zt + ?t rt /3)
||zt+1 ? zt ||
= min 1,
||zt+1 ? x
?t+1 ||
= x
?t+1 + ?t+1 (zt+1 ? x
?t+1 )
Step (1) updates xt according to GIGA?s standard gradient update and stores the result as
x
?t+1 . Step (2) updates zt in the same manner, but with a smaller step-size. Step (3) makes
a final adjustment on xt+1 by moving it toward zt+1 . The magnitude of this adjustment is
limited by the change in zt that occurred in step (2).
A key factor in understanding this algorithm is the observance that a strategy a receives
higher reward than a strategy b if and only if the gradient at a is in the direction of b (i.e.,
rt ? (b ? a) > 0). Therefore, the step (3) adjustment is in the direction of the gradient
if and only if zt received higher reward than xt . Notice also that, as long as xt is not
near the boundary, the change due to step (3) is of lower magnitude than the change due
4
WoLF-IGA may, in fact, be a limited variant of the extragradient method [13] for variational
inequality problems. The extragradient algorithm is guaranteed to converge to a Nash equilibrium
in self-play for all zero-sum games. Like WoLF-IGA, though, it does not have any known regret
guarantees, but more importantly requires the other players? payoffs to be known.
to step (1). Hence, the combination of steps (1) and (3) result in a change with two key
properties. First, the change is in the direction of positive gradient. Second, the magnitude
of the change is larger when zt received higher reward than xt . So, we can interpret the
update rule as a variation on the WoLF principle of ?win or learn fast?, i.e., the algorithm is
learning faster if and only if its strategy x is losing to strategy z. GIGA-WoLF is a major
improvement on the original presentation of WoLF-IGA, where winning was determined
by comparison with an equilibrium strategy that was assumed to be given. Not only is less
knowledge required, but the use of a GIGA-updated strategy to determine winning will
allow us to prove guarantees on the algorithm?s regret.
In the next section we present a theoretical examination of GIGA-WoLF?s regret in nplayer, n-action games, along with a limited guarantee of convergence in two-player, twoaction games. In Section 5, we give experimental results of learning using GIGA-WoLF,
demonstrating that convergence extends beyond the theoretical analysis presented.
4
Theoretical Analysis
We begin by examining GIGA-WoLF?s regret against an unknown opponent strategy. We
will prove the following bound on average regret.
?
Theorem 1 If ?t = 1/ t, the regret of GIGA-WoLF is,
?
?
2
(2 T ? 1).
RT ? 2 T + |A|rmax
Therefore, limT ?? RT /T ? 0, hence GIGA-WoLF has no-regret.
Proof. We begin with a brief overview of the proof. We will find a bound on the regret of
the strategy xt with respect to the dynamic strategy zt . Since zt is unmodified GIGA, we
already have a bound on the regret of zt with respect to any static strategy. Hence, we can
bound the regret of xt with respect to any static strategy.
We start by examining the regret of xt with respect to zt using a similar analysis as used
by Zinkevich [11]. Let ?x?z
refer to the difference in expected payoff between zt and xt
t
at time t, and Rx?z
be the sum of these differences, i.e., the total regret of xt with respect
T
to zt ,
T
X
x?z
x?z
?t
= rt ? (zt ? xt )
RT
?
?x?z
.
t
t=1
We will use the following potential function, ?t ? (xt ? zt )2 /2?t . We can examine how
this potential changes with each step of the update. ??1t , ??2t , and ??3t refers to the
change in potential caused by steps (1), (2), and (3), respectively. ??4t refers to the change
in potential caused by the learning rate change from ?t?1 to ?t . This gives us the following
equations for the potential change.
??1t+1
=
??2t+1
=
??3t+1
=
??4t+1
=
??t+1
=
1/2?t ((?
xt+1 ? zt )2 ? (xt ? zt )2 )
1/2?t ((?
xt+1 ? zt+1 )2 ? (?
xt+1 ? zt )2 )
1/2?t ((xt+1 ? zt+1 )2 ? (?
xt+1 ? zt+1 )2 )
(1/2?t+1 ? 1/2?t ) (xt+1 ? zt+1 )2
??1t+1 + ??2t+1 + ??3t+1 + ??4t+1
Notice that if ?t+1 = 1 then xt+1 = zt+1 . Hence ?t+1 = 0, and ??2t+1 + ??3t+1 ? 0.
If ?t+1 < 1, then ||xt+1 ? x
?t+1 || = ||zt+1 ? zt ||. Notice also that in this case xt+1 is
co-linear and between x
?t+1 and zt+1 . So,
||?
xt+1 ? zt+1 ||
= ||?
xt+1 ? xt+1 || + ||xt+1 ? zt+1 ||
= ||zt+1 ? zt || + ||xt+1 ? zt+1 ||
We can bound the left with the triangle inequality,
||?
xt+1 ? zt+1 ||
||xt+1 ? zt+1 ||
? ||?
xt+1 ? zt || + ||zt ? zt+1 ||
? ||?
xt+1 ? zt ||.
So regardless of ?t+1 , ??2t+1 + ??3t+1 < 0. Hence, ??t+1 ? ??1t+1 + ??4t+1 .
We will now use this bound on the change in the potential to bound the regret of x t with
respect to zt . We know from Zinkevich that,
(?
xt+1 ? zt )2 ? (xt ? zt )2 ? 2?t rt ? (zt ? xt ) + ?t2 rt2 .
Therefore,
?x?z
t
1
(?
xt+1 ? zt )2 ? (xt ? zt )2 ? rt2
2?t
? ???1t+1 + 1/2?t rt2 ? ???t+1 + ??4t+1 + 1/2?t rt2 .
? ?
2
Since we assume rewards are bounded?by rmax we can bound rt2 by |A|rmax
. Summing up
regret and using the fact that ?t = 1/ t, we get the following bound.
Rx?z
T
T
X
?t
2
|A|rmax
2
t=1
T
2
X
1
|A|rmax
? (?1 ? ?T ) +
?1 +
?t
?T
2
t=1
?
?
2
?
T + |A|rmax
( T ? 1/2)
?
???t + ??4t +
We know that GIGA?s regret with respect to any strategy is bounded by the same value (see
Inequality 2). Hence,
?
?
2
(2 T ? 1),
RT ? 2 T + |A|rmax
as claimed.
The second criterion we want to consider is convergence. As with IGA, WoLF-IGA, and
other algorithms, our theoretical analysis will be limited to two-player, two-action generalsum games. We further limit ourselves to the situation of GIGA-WoLF playing ?against?
GIGA. These restrictions are a limitation of the proof method, which uses a case-by-case
analysis that is combinatorially impractical for the case of self-play. This is not necessarily
a limitation on GIGA-WoLF?s convergence. This theorem along with the empirical results
we present later in Section 5 give a strong sense of GIGA-WoLF?s convergence properties.
The full proof can be found in [1].
Theorem 2 In a two-player, two-action repeated game, if one player follows the GIGAWoLF algorithm and the other follows the GIGA algorithm, then their strategies will converge to a Nash equilibrium.
5
Experimental Analysis
We have presented here two theoretical properties of GIGA-WoLF relating to guarantees
on both regret and convergence. There have also been extensive experimental results performed on GIGA-WoLF in a variety of normal-form games [1]. We summarize the results
here. The purpose of these experiments was to demonstrate the theoretical results from the
previous section as well as explore the extent to which the results (convergence, in particular) can be generalized. In that vein, we examined the same suite of normal-form games
used in experiments with WoLF-PHC, the practical variant of WoLF-IGA [7].
One of the requirements of GIGA-WoLF (and GIGA) is knowledge of the entire reward
vector (rt ), which requires knowledge of the game and observation of the opponent?s action. In practical situations, one or both of these are unlikely to be available. Instead, only
the reward of the selected action is likely to be observable. We have relaxed this requirement in these experiments by providing GIGA-WoLF (and GIGA) with only estimates of
the gradient from stochastic approximation. In particular, after selecting action a and receiving reward r?a , we update the current estimate of action a?s component of the reward
vector, rt+1 = rt + ?t (?
ra ? 1a ? rt )1a , where ?t is the learning rate. This is a standard
method of estimation commonly used in reinforcement learning (e.g., Q-learning).
For almost all of the games explored, including two-player, two-action games as well as
n-action zero-sum games, GIGA-WoLF strategies converged in self-play to equilibrium
strategies of the game. GIGA?s strategies, on the other hand, failed to converge in self-play
over the same suite of games. These results are nearly identical to the PHC and WoLF-PHC
experiments over the same games. A prototypical example of these results is provided in
Figure 1(a) and (b), showing strategy trajectories while learning in Rock-Paper-Scissors.
GIGA?s strategies do not converge, while GIGA-WoLF?s strategies do converge. GIGAWoLF also played directly against GIGA in this game resulting in convergence, but with
a curious twist. The resulting expected and average payoffs are shown in Figure 1(c).
Since both are no-regret learners, average payoffs are guaranteed to go to zero, but the
short-term payoff is highly favoring GIGA-WoLF. This result raises an interesting question
about the relative short-term performance of no-regret learning algorithms, which needs to
be explored further.
1
0.4
0.2
Reward
0.6
0.6
0.4
0
0
0.2 0.4 0.6 0.8
Pr(Rock)
(a) GIGA
1
0.04
0
-0.04
0.2
0
Average Reward
Expected Reward
0.08
0.8
Pr(Paper)
0.8
Pr(Paper)
0.12
1
-0.08
0
0.2 0.4 0.6 0.8
Pr(Rock)
(b) GIGA-WoLF
1
0
500000
Iterations
1e+06
(c) GIGA v. GIGA-WoLF
Figure 1: Trajectories of joint strategies in Rock-Paper-Scissors when both players use
GIGA (a) or GIGA-WoLF (b). Also shown (c) are the expected and average payoffs of the
players when GIGA and GIGA-WoLF play against each other.
GIGA-WoLF did not lead to convergence in all of the explored games. The ?problematic?
Shapley?s game, for which many similarly convergent algorithms fail in, also resulted in
non-convergence for GIGA-WoLF. On the other hand, this game has the interesting property that both players? when using GIGA-WoLF (or GIGA) actually achieve negative regret.
In other words, the algorithms are outperforming any static strategy to which they could
converge upon. This suggests a new desirable property for future multiagent (or online)
learning algorithms, negative non-convergence regret (NNR). An algorithm has NNR, if it
satisfies the no-regret property and either (i) achieves negative regret or (ii) its strategy converges. This property combines the criteria of regret and convergence, and GIGA-WoLF is
a natural candidate for achieving this compelling result.
6
Conclusion
We introduced GIGA-WoLF, a new gradient-based algorithm, that we believe is the first
to simultaneously address two criteria: no-regret and convergence. We proved GIGAWoLF has no-regret. We also proved that in a small class of normal-form games, GIGAWoLF?s strategy when played against GIGA will converge to a Nash equilibrium. We
summarized experimental results of GIGA-WoLF playing in a variety of zero-sum and
general-sum games. These experiments verified our theoretical results and exposed two
interesting phenomenon that deserve further study: short-term performance of no-regret
learners and the new desiderata of negative non-convergence regret. We expect GIGAWoLF and these results to be the foundation for further understanding of the connections
between the regret and convergence criteria.
References
[1] Michael Bowling. Convergence and no-regret in multiagent learning. Technical Report TR04-11, Department of Computing Science, University of Alberta, 2004.
[2] Michael L. Littman. Markov games as a framework for multi-agent reinforcement
learning. In Proceedings of the Eleventh International Conference on Machine Learning, pages 157?163, 1994.
[3] Junling Hu and Michael P. Wellman. Multiagent reinforcement learning: Theoretical
framework and an algorithm. In Proceedings of the Fifteenth International Conference on Machine Learning, pages 242?250, 1998.
[4] Amy Greenwald and Keith Hall. Correlated Q-learning. In Proceedings of the AAAI
Spring Symposium Workshop on Collaborative Learning Agents, 2002.
[5] Caroline Claus and Craig Boutilier. The dynamics of reinforcement learning in cooperative multiagent systems. In Proceedings of the Fifteenth National Conference on
Artificial Intelligence, pages 746?752, 1998.
[6] Satinder Singh, Michael Kearns, and Yishay Mansour. Nash convergence of gradient dynamics in general-sum games. In Proceedings of the Sixteenth Conference on
Uncertainty in Artificial Intelligence, pages 541?548, 2000.
[7] Michael Bowling and Manuela Veloso. Multiagent learning using a variable learning
rate. Artificial Intelligence, 136:215?250, 2002.
[8] Yu-Han Chang and Leslie Pack Kaelbling. Playing is believing: the role of beliefs
in multi-agent learning. In Advances in Neural Information Processing Systems 14,
2001.
[9] Sergiu Hart and Andreu Mas-Colell. A simple adaptive procedure leading to correlated equilibrium. Econometrica, 68:1127?1150, 2000.
[10] Peter Auer, Nicol`o Cesa-Bianchi, Yoav Freund, and Robert E. Schapire. Gambling in
a rigged casino: The adversarial multi-arm bandit problem. In 36th Annual Symposium on Foundations of Computer Science, pages 322?331, 1995.
[11] Martin Zinkevich. Online convex programming and generalized infinitesimal gradient ascent. In Proceedings of the Twentieth International Conference on Machine
Learning, pages 928?925, 2003.
[12] Amir Jafari, Amy Greenwald, David Gondek, and Gunes Ercal. On no-regret learning,
fictitious play, and nash equilibrium. In Proceedings of the Eighteenth International
Conference on Machine Learning, pages 226?223, 2001.
[13] G. M. Korpelevich. The extragradient method for finding saddle points and other
problems. Matecon, 12:747?756, 1976.
| 2673 |@word exploitation:1 version:1 stronger:1 norm:1 rigged:1 hu:1 seek:2 exclusively:1 selecting:2 past:1 current:2 update:9 stationary:4 intelligence:3 selected:3 amir:1 short:3 along:2 symposium:2 prove:4 shapley:1 combine:1 eleventh:1 manner:1 introduce:1 theoretically:1 x0:2 ra:1 expected:11 behavior:1 roughly:1 examine:4 nor:1 multi:3 alberta:3 actual:1 considering:1 project:1 begin:2 bounded:4 provided:1 argmin:1 rmax:10 finding:2 impractical:1 suite:2 guarantee:9 subclass:1 gondek:1 tricky:1 twoaction:1 internally:1 before:1 positive:1 limit:2 lure:1 deceptive:1 dynamically:1 examined:3 challenging:1 suggests:1 co:1 limited:5 practical:2 regret:57 procedure:1 universal:1 empirical:4 convenient:1 matching:2 word:4 refers:3 suggest:1 get:1 cannot:2 clever:1 selection:1 context:1 applying:1 optimize:1 zinkevich:4 restriction:1 demonstrated:1 eighteenth:1 maximizing:1 go:1 attention:1 regardless:1 independently:1 convex:1 assigns:1 amy:2 insight:1 avoidance:1 deceptively:1 rule:2 importantly:1 ercal:1 his:1 notion:1 nnr:3 variation:1 updated:2 target:2 play:14 yishay:1 ualberta:1 losing:1 programming:1 us:1 cooperative:1 vein:1 role:1 worst:1 e8:1 environment:3 nash:9 reward:25 littman:1 econometrica:1 dynamic:6 depend:1 singh:2 raise:1 exposed:1 upon:2 learner:11 completely:1 triangle:1 joint:3 iga:11 fast:2 artificial:3 outcome:2 larger:2 say:1 otherwise:1 final:1 online:3 rock:5 combining:1 achieve:2 sixteenth:1 description:1 convergence:40 requirement:2 r1:6 guaranteeing:1 converges:5 wider:1 depending:1 rt2:5 received:2 keith:1 strong:2 c:1 implies:2 direction:3 safe:1 stochastic:1 hall:1 normal:9 equilibrium:17 mapping:1 claim:1 major:1 vary:1 sought:1 a2:1 achieves:3 purpose:1 estimation:1 combinatorially:1 always:1 desirability:1 focus:2 improvement:1 believing:1 adversarial:1 sense:1 entire:1 unlikely:1 bandit:1 favoring:1 equal:1 identical:2 yu:1 nearly:2 future:1 simplex:1 t2:1 report:1 oblivious:1 simultaneously:4 resulted:1 national:1 ourselves:1 ab:1 stationarity:2 highly:1 evaluation:6 analyzed:1 wellman:1 poorer:1 tuple:1 unless:1 theoretical:10 compelling:1 measuring:1 unmodified:1 yoav:1 leslie:1 contested:1 introducing:1 kaelbling:2 examining:2 colell:2 chooses:1 international:4 receiving:2 michael:6 aaai:1 cesa:1 choose:1 worse:1 leading:1 potential:6 summarized:1 casino:1 notable:1 caused:2 scissors:3 later:1 performed:1 analyze:1 start:1 collaborative:1 weak:1 craig:1 trajectory:3 rx:2 history:1 converged:1 caroline:1 infinitesimal:2 against:9 proof:4 static:7 proved:4 knowledge:3 actually:1 back:1 auer:1 focusing:1 normalform:2 higher:3 response:5 though:3 just:1 lastly:1 hand:3 receives:2 believe:1 true:1 hence:6 bowling:4 game:46 self:10 criterion:15 generalized:3 demonstrate:1 performs:1 variational:1 recently:1 common:4 overview:1 twist:1 he:2 occurred:1 relating:1 interpret:1 refer:1 ai:1 unconstrained:1 consistency:1 similarly:1 had:1 moving:1 han:1 longer:1 korpelevich:1 recent:2 store:1 certain:1 claimed:1 inequality:3 outperforming:1 arbitrarily:1 continue:1 exploited:4 relaxed:1 converge:13 determine:1 ii:1 multiple:1 full:1 desirable:1 technical:1 faster:1 veloso:1 offer:1 long:1 hart:2 award:1 a1:9 converging:1 variant:2 desideratum:1 circumstance:1 expectation:2 fifteenth:2 iteration:2 represent:1 limt:3 achieved:1 receive:1 addition:1 want:1 ascent:8 claus:1 call:1 curious:1 near:1 variety:4 switch:1 isolation:2 suboptimal:1 whether:2 peter:1 action:28 repeatedly:1 boutilier:1 simplest:1 schapire:1 problematic:1 notice:3 track:1 write:1 key:5 demonstrating:3 achieving:1 changing:1 neither:1 verified:1 sum:7 uncertainty:1 extends:1 almost:2 reasonable:1 sergiu:1 bound:9 followed:1 distinguish:1 guaranteed:3 played:2 convergent:1 identifiable:1 annual:1 strength:1 ri:1 min:1 concluding:1 spring:1 performing:1 martin:1 department:2 according:2 combination:1 smaller:1 restricted:1 pr:4 legal:1 equation:1 fail:2 know:3 available:2 opponent:9 away:1 undermining:1 stepsize:1 alternative:1 original:1 include:1 exploit:1 already:1 question:1 looked:1 strategy:51 rt:24 traditional:1 gradient:18 win:2 extent:1 toward:1 gunes:1 providing:1 difficult:1 susceptible:1 robert:1 debate:1 negative:5 stated:1 zt:49 policy:2 unknown:2 bianchi:1 observation:1 markov:1 situation:4 extended:3 payoff:11 andreu:1 mansour:2 canada:1 overcoming:1 introduced:1 david:1 namely:1 required:1 extensive:1 connection:1 address:2 able:1 beyond:2 deserve:1 usually:1 challenge:7 summarize:1 gaining:1 including:4 max:1 belief:1 natural:1 examination:1 generalsum:1 arm:1 brief:2 numerous:1 literature:1 l2:1 understanding:2 nicol:1 asymptotic:1 relative:1 freund:1 multiagent:13 loss:2 expect:1 prototypical:1 limitation:2 interesting:3 fictitious:1 foundation:2 agent:15 t6g:1 principle:2 playing:3 formal:1 weaker:1 allow:1 penny:2 boundary:1 evaluating:1 commonly:1 reinforcement:5 adaptive:1 observable:1 keep:1 satinder:1 summing:1 manuela:1 assumed:1 table:3 learn:5 pack:1 ca:1 correlated:5 necessarily:2 did:1 jafari:1 proliferate:1 repeated:1 gambling:1 edmonton:1 winning:2 candidate:1 third:1 theorem:3 xt:50 showing:1 explored:5 r2:3 consist:1 workshop:1 magnitude:3 simply:1 explore:2 likely:1 twentieth:1 saddle:1 failed:1 adjustment:3 chang:2 wolf:42 corresponds:1 satisfies:1 ma:2 goal:3 presentation:1 greenwald:2 change:14 deception:1 included:1 experimentally:1 determined:1 extragradient:3 kearns:2 total:7 experimental:6 player:25 rarely:1 select:1 exception:1 giga:54 phc:4 phenomenon:1 handling:1 |
1,843 | 2,674 | Maximising Sensitivity in a Spiking Network
Anthony J. Bell,
Redwood Neuroscience Institute
1010 El Camino Real, Suite 380
Menlo Park, CA 94025
[email protected]
Lucas C. Parra
Biomedical Engineering Department
City College of New York
New York, NY 10033
[email protected]
Abstract
We use unsupervised probabilistic machine learning ideas to try to explain the kinds of learning observed in real neurons, the goal being
to connect abstract principles of self-organisation to known biophysical processes. For example, we would like to explain Spike TimingDependent Plasticity (see [5,6] and Figure 3A), in terms of information
theory. Starting out, we explore the optimisation of a network sensitivity measure related to maximising the mutual information between input
spike timings and output spike timings. Our derivations are analogous to
those in ICA, except that the sensitivity of output timings to input timings is maximised, rather than the sensitivity of output ?firing rates? to
inputs. ICA and related approaches have been successful in explaining
the learning of many properties of early visual receptive fields in rate coding models, and we are hoping for similar gains in understanding of spike
coding in networks, and how this is supported, in principled probabilistic
ways, by cellular biophysical processes. For now, in our initial simulations, we show that our derived rule can learn synaptic weights which can
unmix, or demultiplex, mixed spike trains. That is, it can recover independent point processes embedded in distributed correlated input spike
trains, using an adaptive single-layer feedforward spiking network.
1
Maximising Sensitivity.
In this section, we will follow the structure of the ICA derivation [4] in developing the
spiking theory. We cannot claim, as before, that this gives us an information maximisation
algorithm, for reasons that we will delay addressing until Section 3. But for now, to first
develop our approach, we will explore an interim objective function called sensitivity which
we define as the log Jacobian of how input spike timings affect output spike timings.
1.1
How to maximise the effect of one spike timing on another.
Consider a spike in neuron j at time tl that has an effect on the timing of another spike in
neuron i at time tk . The neurons are connected by a weight wij . We use i and j to index
neurons, and k and l to index spikes, but sometimes for convenience we will use spike
indices in place of neuron indices. For example, wkl , the weight between an input spike l
and an output spike k, is naturally understood to be just the corresponding wij .
dtl
threshold potential
dtk
du
u(t)
R(t)
resting potential
tk
output spikes
tl
input spikes
Figure 1: Firing time tk is determined by the time of threshold crossing. A change of an
input spike time dtl affects, via a change of the membrane potential du the time of the
output spike by dtk .
In the simplest version of the Spike Response Model [7], spike l has an effect on spike k that
depends on the time-course of the evoked EPSP or IPSP, which we write as R kl (tk ? tl ).
In general, this Rkl models both synaptic and dendritic linear responses to an input spike,
and thus models synapse type and location. For learning, we need only consider the value
of this function when an output spike, k, occurs.
In this model, depicted in Figure 1, a neuron adds up its spiking inputs until its membrane potential, ui (t), reaches threshold at time tk . This threshold we will often, again for
convenience, write as uk ? ui (tk , {tl }), and it is given by a sum over spikes l:
X
uk =
wkl Rkl (tk ? tl ) .
(1)
l
To maximise timing sensitivity, we need to determine the effect of a small change in the
input firing time tl on the output firing time tk . (A related problem is tackled in [2].) When
tl is changed by a small amount dtl the membrane potential will change as a result. This
change in the membrane potential leads to a change in the time of threshold crossing dt k .
The contribution to the membrane potential, du, due to dtl is (?uk /?tl )dtl , and the change
in du corresponding to a change dtk is (?uk /?tk )dtk . We can relate these two effects
by noting that the total change of the membrane potential du has to vanish because u k is
defined as the potential at threshold. ie:
du =
?uk
?uk
dtk +
dtl = 0 .
?tk
?tl
(2)
This is the total differential of the function uk = u(tk , {tl }), and is a special case of the
implicit function theorem. Rearranging this:
dtk
?uk
?uk
=?
= ?wkl R? kl /u? k .
(3)
dtl
?tl
?tk
Now, to connect with the standard ICA
P derivation [4], recall the ?rate? (or sigmoidal) neuron, for which yi = gi (ui ) and ui = j wij xj . For this neuron, the output dependence on
input is ?yi /?xj = wij gi0 while the learning gradient is:
?yi
?
= 1 ? fi (ui )xj
log
?wij
?xj wij
(4)
The analogous learning gradient for the spiking case, from (3), is:
P
?
dtk
?
1
a j(a)Rka
log
?
.
=
?wij
dtl wij
u? k
(5)
where the ?score functions?, fi , are defined in terms of a density estimate on the summed
?
?
inputs: fi (ui ) = ?u
log gi0 = ?u
log p?(ui ).
i
i
where j(a) = 1 if spike a came from neuron j, and 0 otherwise.
Comparing the two cases in (4) and (5), we see that the input variable xj has become
the temporal derivative of the sum of the EPSPs coming from synapse j, and the output
variable (or score function) fi (ui ) has become u? ?1
k , the inverse of the temporal derivative
of the membrane potential at threshold. It is intriguing (A) to see this quantity appear as
analogous to the score function in the ICA likelihood model, and, (B) to speculate that
experiments could show that this? voltage slope at threshold? is a hidden factor in STDP
data, explaining some of the scatter in Figure 3A. In other words, an STDP datapoint should
lie on a 2-surface in a 3D space of {?w, ?t, u? k }. Incidentally, u? k shows up in any
learning rule optimising an objective function involving output spike timings.
1.2
How to maximise the effect of N spike timings on N other ones.
Now we deal with the case of a ?square? single-layer feedforward mapping between spike
timings. There can be several input and output neurons, but here we ignore which neurons
are spiking, and just look at how the input timings affect the output timings. This is captured
in a Jacobian matrix of all timing dependencies we call T. The entries of this matrix are
Tkl ? ?tk /?tl . A multivariate version of the sensitivity measure introduced in the previous
section is the log of the absolute determinant of the timing matrix, ie: log |T|. The full
derivation for the gradient ?W log |T| is in the Appendix. Here, we again draw out the
analogy between Square ICA [4] and this gradient, as follows. Square ICA with a network
y = g(Wx) is:
?W ? ?W log |J| = W?1 ? f (u)xT
(6)
where the Jacobian J has entries ?yi /?xj and the score functions are now, fi (u) =
Q
?
? ?u
log p?(u) for the general likelihood case, with p?(u) = i gi0 being the special case of
i
ICA. We will now split the gradient in (6) according to the chain rule:
?W log |J|
=
[?J log |J|] ? [?W J]
?T
j(l)
= J
? Jkl i(k)
? fk (u)xj
.
wkl
(7)
[?T log |T|] ? [?W T]
!#
"
P
?
?T
j(l)
a j(a)Rka
= T
? Tkl i(k)
?
wkl
u? k
(9)
(8)
In this equation, i(k) = ?ik and j(l) = ?jl . P
The righthand term is a 4-tensor with entries
?Jkl /?wij , and ? is defined as A ? Bij = kl Akl Bklij . We write the gradient this way
to preserve, in the second term, the independent structure of the 1 ? 1 gradient term in
(4), and to separate a difficult derivation into two easy parts. The structure of (8) holds up
when we move to the spiking case, giving:
?W log |T| =
(10)
where i(k) is now defined as being 1 if spike k occured in neuron i, and 0 otherwise. j(l)
and j(a) are analogously defined.
Because the T matrix is much bigger than the J matrix, and because it?s entries are more
complex, here the similarity ends. When (10) is evaluated for a single weight influencing a
single spike coupling (see the Appendix for the full derivation), it yields:
?wkl ?
? log |T|
Tkl ?1
=
T lk ? 1 ,
?wkl
wkl
(11)
This is a non-local update involving a matrix inverse at each step. In the ICA case of (6),
such an inverse was removed by the Natural Gradient transform (see [1]), but in the spike
timing case, this has turned out not to be possible, because of the additional asymmetry
introduced into the T matrix (as opposed to the J matrix) by the R? kl term in (3).
2
Results.
Nonetheless, this learning rule can be simulated. It requires running the network for a
while to generate spikes (and a corresponding T matrix), and then for each input/output
spike coupling, the corresponding synapse is updated according to (11). When this is done,
and the weights learn, it is clear that something has been sacrificed by ignoring the issue
of which neurons are producing the spikes. Specifically, the network will often put all the
output spikes on one output neuron, with the rates of the others falling to zero. It is happy
to do this, if a large log |T| can thereby be achieved, because we have not included this
?which neuron? information in the objective. We will address these and other problems in
Section 3, but now we report on our simulation results on demultiplexing.
2.1
Demultiplexing spike trains.
An interesting possibility in the brain is that ?patterns? are embedded in spatially distributed
spike timings that are input to neurons. Several patterns could be embedded in single input
trains. This is called multiplexing. To extract and propagate these patterns, the neurons
must demultiplex these inputs using its threshold nonlinearity. Demultiplexing is the ?point
process? analog of the unmixing of independent inputs in ICA. We have been able to robustly achieve demultiplexing, as we now report.
We simulated a feed-forward network with 3 integrate-and-fire neurons and inputs from 3
presynaptic neurons. Learning followed (11) where we replace the inverse by the pseudoinverse computed on the spikes generated during 0.5 s. The pseudo-inverse is necessary
because even though on average, the learning matches number of output spikes to number
of input spikes, the matrix T is still not usually square and so its actual inverse cannot be
taken.
In addition, in these simulations, an additional term is introduced in the learning to make
sure all the output neurons fire with equal probability. This partially counters the ignoral of
the ?which neuron? information, which we explained above. Assuming Poisson spike count
ni for the ith output neuron with equalP
firing rate n
? i it is easy to derive in an approximate
term that will control the spike count, i (?
ni ? ni ). The target firing rates n
? i were set to
match the ?source? spike train in this example.
The network learns to demultiplex mixed spike trains, as shown in Figure 2. This demultiplexing is a robust property of learning using (11) with this new spike-controlling term.
Finally, what about the spike-timing dependendence of the observed learning? Does it
match experimental results? The comparison is made in Figure 3, and the answer is no.
There is a timing-dependent transition between depression and potentiation in our result
Spike Trains
mixing
mixed input trains
1
1
0.8
2
0.6
3
0
50
100
150
200
250
300
350
400
450
0.4
500
0.2
output
1
0
2
3
synaptic weights
0
50
100
150
200
250
300
350
400
450
500
original spike train
1
1
0.5
2
0
3
0
50
100
150
200
250
time in ms
300
350
400
450
500
?0.5
Figure 2: Unmixed spike trains. The input (top lef) are 3 spike trains which are a mixture
of three independent Poison processes (bottom left). The network unmixes the spike train
to approximately recover the original (center left). In this example 19 spikes correspond to
the original with 4 deletion and 2 insertions. The two panels at the right show the mixing
(top) and synaptic weight matrix after training (bottom).
in Figure 3B, but it is not a sharp transition like the experimental result in Figure 3A. In
addition, it does not transition at zero (ie: when tk ? tl = 0), but at a time offset by the
rise time of the EPSPs. In earlier experiments, in which we tranformed the gradient in
(11) by an approximate inverse Hessian, to get an approximate Natural Gradient method,
a sharp transition did emerge in simulations. However, the approximate inverse Hessian
was singular, and we had to de-emphasise this result. It does suggest, however, that if the
Natural Gradient transform can be usefully done on some variant of this learning rule, it
may well be what accounts for the sharp transition effect of STDP.
3
Discussion
Although these derivations started out smoothly, the reader possibly shares the authors?
frustration at the approximations involved here. Why isn?t this simple, like ICA? Why
don?t we just have a nice maximum spikelihood model, ie: a density estimation algorithm
for multivariate point processes, as ICA was a model in continuous space? We are going
to be explicit about the problems now, and will propose a direction where the solution may
lie. The over-riding problem is: we are unable to claim that in maximising log |T|, we are
maximising the mutual information between inputs and outputs because:
1. The Invertability Problem. Algorithms such as ICA which maximise log Jacobians can
only be called Infomax algorithms if the network transformation is both deterministic and
invertable. The Spike Response Model is deterministic, but it is not invertable in general.
When not invertable, the key formula (considering here vectors of input and output timings,
tin and tout )is transformed from simple to complex. ie:
Z
p(tin )
p(tin )
p(tout ) =
becomes p(tout ) =
d tin
(12)
|T|
solns tin |T|
Thus when not invertable, we need to know the Jacobians of all the inputs that could have
caused an output (called here ?solns?), something we simply don?t know.
2. The ?Which Neuron? Problem. Instead of maximising the mutual information
I(tout , tin ), we should be maximising I(tiout , tiin ), where the vector ti is the timing
(B) Gradient
150
100
100
? w (a.u.)
? w / w (%)
(A) STDP
150
50
0
?50
?100
?100
50
0
?50
?50
0
? t (ms)
50
100
?100
?20
0
20
40
60
? t (ms)
80
100
Figure 3: Dependence of synaptic modification on pre/post inter-spike interval. Left (A):
From Froemke & Dan, Nature (2002)]. Dependence of synaptic modification on pre/post
inter-spike interval in cat L2/3 visual cortical pyramidal cells in slice. Naturalistic spike
trains. Each point represents one experiment. Right (B): According to Equation (11).
Each point corresponds to an spike pair between approximately 100 input and 100 output
spikes.
vector, t, with the vector, i, of corresponding neuron indices, concatenated. Thus, ?who
spiked?? should be included in the analysis as it is part of the information.
3. The Predictive Information Problem. In ICA, since there was no time involved, we did
not have to worry about mutual informations over time between inputs and outputs. But in
the spiking model, output spikes may well have (predictive) mutual information with future
input spikes, as well as the usual (causal) mutual information with past input spikes. The
former has been entirely missing from our analysis so far.
These temporal and spatial infomation dependencies missing in our analysis so far, are
thrown into a different light by a single empirical observation, which is that Spike TimingDependent Plasticity is not just a feedforward computation like the Spike Response Model.
Specifically, there must be at least a statistical, if not a causal, relation between a real
synapse?s plasticity and its neuron?s output spike timings, for Figure 3B to look like it
does.
It seems we have to confront the need for both a ?memory? (or reconstruction) model, such
as the T we have thus far dealt with, in which output spikes talk about past inputs, and a
?prediction? model, in which they talk about future inputs. This is most easily understood
from the point of view of Barber & Agakov?s variational Infomax algorithm [3]. They
argue for optimising a lower bound on mutual information, which, for our neurons?, would
be expressed using an inverse model p?, as follows:
e in , tiout ) = H(tiin ) ? hlog p?(tiin |tiout )ip(ti ,ti ) ? I(tiin , tiout )
I(ti
in
out
(13)
In a feedforward model, H(tiin ) may be disregarded in taking gradients, leading us to the
optimisation of a ?memory-prediction? model p?(tiin |tiout ) related to something supposedly happening in dendrites, somas and at synapses. In trying to guess what this might be,
it would be nice if the math worked out. We need a square Jacobian matrix, T, so that
|T| = p?(tiin |tiout ) can be our memory/prediction model. Now let?s rename our feedfor?
?
ward timing Jacobian T (?up the dendritic trees?), as T, and let?s fantasise that there is
?
?
some, as yet unspecified, feedback Jacobian T (?down the dendritic trees?), which covers
?
?
electrotonic influences as they spread from soma to synapse, and which T can be combined with by some operation ??? to make things square. Imagine further, that doing this
?
? ?
?
yields a memory/prediction model on the inputs. Then the T we are looking for is T ? T,
? ?
?
?
and the memory-prediction model is: p?(tiin |tiout ) = T ? T
?
?
?
?
Ideally, the entries of T should be as before, ie: T kl = ?tk /?tl . What should the entries
?
?
?
?
?
?
of T be? Becoming just one step more concrete, suppose T had entries T lk = ?cl /?tk ,
where cl is some, as yet unspecified, value, or process, occuring at an input synapse when
spike l comes in. What seems clear is that ? should combine the correctly tensorised forms
?
?
?
?
?
? ?
?
of T and T (giving them each 4 indices ijkl), so that T = T ? T sums over the spikes
k and l to give a I ? J matrix, where I is the number of output neurons, and J the number
of input neurons. Then our quantity, T, would represent all dependencies of input neuronal
activity on output activity, summed over spikes.
?
?
Further, we imagine that T contains reverse (feedback) electrotonic transforms from soma
?
?
to synapse R lk that are somehow symmetrically related to the feedforward Spike Re?
?
sponses from synapse to soma, which we now rename R kl . Thinking for a moment in terms
of somatic k and synaptic l, voltages V , currents I and linear cable theory, the synapse to
?
?
?
?
soma transform, R kl would be related to an impedance in Vk = Il Z kl , while the soma
?
?
?
?
to synapse transform, R lk would be related to an admittance in Il = Vk Y lk [8]. The
?
?
?
?
symmetry in these equations is that Z kl is just the inverse conjugate of Y lk .
Finally, then, what is cl ? And what is its relation to the calcium concentration, [Ca2+ ]l ,
at a synapse, when spike l comes in? These questions naturally follow from considering
the experimental data, since it is known that the calcium level at synapses is the critical
integrating factor in determining whether potentiation or depression occurs [5].
4
Appendix: Gradient of log |T| for the full Spike Response Model.
Here we give full details of the gradient for Gerstner?s Spike Response Model [7]. This is
a general model for which Integrate-and-Fire is a special case. In this model the effect of
a presynaptic spike at time tl on the membrane potential at time t is described by a post
synaptic potential or spike response, which may also depend on the time that has passed
since the last output spike tk?1 , hence the spike response is written as R(t ? tk?1 , t ? tl ).
This response is weighted by the synaptic strength wl . Excitatory or inhibitory synapses are
determined by the sign of wl . Refractoriness is incorporated by adding a hyper-polarizing
contribution (spike-afterpotential) to the membrane potential in response to the last preceding spike ?(t ? tk?1 ). The membrane potential as a function of time is therefore given
by
X
u(t) = ?(t ? tk?1 ) +
wl R(t ? tk?1 , t ? tl ) .
(14)
l
We have ignored here potential contributions from external currents which can easily be included without modifying the following derivations. The output firing times t k are defined
as the times for which u(t) reaches firing threshold from below. We consider a dynamic
threshold, ?(t ? tk?1 ), which may depend on the time since that last spike tk?1 , together
then output spike times are defined implicitly by:
t = tk : u(t) = ?(t ? tk?1 ) and
du(t)
> 0.
dt
For this more general model Tkl is given by
?1
? k ? tk?1 , tk ? tl , )
dtk
?u
??
?u
wkl R(t
Tkl =
=?
?
=
,
? k ? tk?1 )
dtl
?tk
?tk
?tl
u(t
? k ) ? ?(t
(15)
(16)
?
? t), u(t),
where R(s,
?
and ?(t)
are derivatives with respect to t. The dependence of Tkl on
tk?1 should be implicitly assumed. It has been omitted to simplify the notation.
Now we compute the derivative of log |T| with respect to wkl . For any matrix T we have
? log |T|/?Tab = [T?1 ]ba . Therefore:
? log |T| X ? log |T| ?Tab X ?1 ?Tab
=
[T ]ba
.
(17)
?wkl
?Tab ?wkl
?wkl
ab
ab
Utilising the Kronecker delta ?ab = (1 if a = b, else 0), the derivative of (16) with respect
to wkl gives:
#
"
? a ? ta?1 , ta ? tb )
?Tab
?
wab R(t
=
P
? a ? ta?1 )
? a ? ta?1 , ta ? tc ) ? ?(t
?wkl
?wkl ?(ta ? ta?1 ) + c wac R(t
? a ? ta?1 , ta ? tb )
R(t
= ?ak ?bl
? a ? ta?1 )
u(t
? a ) ? ?(t
? a ? ta?1 , ta ? tb )?ak R(t
? a ? ta?1 , ta ? tl )
wab R(t
2
? a ? ta?1 )
u(t
? a ) ? ?(t
Tal
?bl
?
.
= ?ak Tab
wab
wal
?
Therefore:
? log |T|
?wkl
=
X
[T?1 ]ba ?ak Tab
ab
=
Tkl
wkl
[T?1 ]lk ?
?bl
Tal
?
wab
wal
X
b
[T?1 ]bk Tkl
!
(18)
(19)
=
Tkl
[T?1 ]lk ? 1 .
wkl
(20)
Acknowledgments
We are grateful for inspirational discussions with Nihat Ay, Michael Eisele, Hong Hui Yu,
Jim Crutchfield, Jeff Beck, Surya Ganguli, Sophi`e Deneve, David Barber, Fabian Theis,
Tony Zador and Arunava Banerjee. AJB thanks all RNI colleagues for many such discussions.
References
[1] Amari S-I. 1997. Natural gradient works efficiently in learning, Neural Computation,
10, 251-276
[2] Banerjee A. 2001. On the Phase-Space Dynamics of Systems of Spiking Neurons.
Neural Computation, 13, 161-225
[3] Barber D. & Agakov F. 2003. The IM Algorithm: A Variational Approach to Information Maximization. Advances in Neural Information Processing Systems 16, MIT Press.
[4] Bell A.J. & Sejnowski T.J. 1995. An information maximization approach to blind
separation and blind deconvolution, Neural Computation, 7, 1129-1159
[5] Dan Y. & Poo M-m. 2004. Spike timing-dependent plasticity of neural circuits, Neuron,
44, 23-30
[6] Froemke R.C. & Dan Y. 2002. Spike-timing-dependent synaptic modification induced
by natural spike trains. Nature, 28, 416: 433-8
[7] Gerstner W. & Kistner W.M. 2002. Spiking neuron models, Camb. Univ. Press
[8] Zador A.M., Agmon-Snir H. & Segev I. 1995. The morphoelectrotonic transform: a
graphical approach to dendritic function, J. Neurosci., 15(3): 1669-1682
| 2674 |@word nihat:1 determinant:1 version:2 dtk:8 seems:2 simulation:4 propagate:1 thereby:1 moment:1 initial:1 contains:1 score:4 past:2 current:2 comparing:1 scatter:1 intriguing:1 jkl:2 must:2 yet:2 written:1 wx:1 plasticity:4 hoping:1 update:1 guess:1 maximised:1 ith:1 unmixed:1 math:1 location:1 org:1 sigmoidal:1 differential:1 become:2 ik:1 dan:3 combine:1 poison:1 inter:2 ica:14 brain:1 ccny:1 actual:1 considering:2 becomes:1 notation:1 panel:1 circuit:1 what:7 kind:1 unspecified:2 akl:1 transformation:1 suite:1 temporal:3 pseudo:1 ti:4 usefully:1 uk:9 control:1 appear:1 producing:1 before:2 maximise:4 engineering:1 timing:26 understood:2 influencing:1 local:1 ak:4 firing:8 becoming:1 approximately:2 might:1 invertability:1 evoked:1 acknowledgment:1 maximisation:1 empirical:1 bell:2 word:1 pre:2 integrating:1 suggest:1 get:1 cannot:2 convenience:2 dtl:9 naturalistic:1 put:1 influence:1 deterministic:2 center:1 missing:2 poo:1 starting:1 zador:2 rule:5 analogous:3 updated:1 target:1 controlling:1 imagine:2 suppose:1 crossing:2 agakov:2 observed:2 bottom:2 connected:1 counter:1 removed:1 principled:1 supposedly:1 ui:8 insertion:1 ideally:1 dynamic:2 depend:2 grateful:1 predictive:2 easily:2 cat:1 talk:2 derivation:8 train:14 sacrificed:1 univ:1 sejnowski:1 hyper:1 otherwise:2 amari:1 gi:1 ward:1 transform:5 ip:1 biophysical:2 propose:1 reconstruction:1 coming:1 epsp:1 turned:1 mixing:2 achieve:1 tout:4 asymmetry:1 unmixing:1 incidentally:1 tk:31 coupling:2 develop:1 derive:1 epsps:2 come:2 direction:1 modifying:1 potentiation:2 dendritic:4 parra:2 im:1 hold:1 stdp:4 mapping:1 claim:2 early:1 omitted:1 agmon:1 estimation:1 wl:3 city:1 weighted:1 mit:1 rather:1 voltage:2 derived:1 vk:2 likelihood:2 ganguli:1 dependent:3 el:1 ajb:1 hidden:1 relation:2 wij:9 going:1 transformed:1 issue:1 lucas:1 spatial:1 special:3 summed:2 mutual:7 tkl:9 field:1 equal:1 optimising:2 represents:1 park:1 look:2 unsupervised:1 yu:1 thinking:1 future:2 others:1 report:2 simplify:1 preserve:1 beck:1 phase:1 fire:3 ab:4 thrown:1 possibility:1 righthand:1 mixture:1 light:1 chain:1 necessary:1 snir:1 tree:2 re:1 causal:2 arunava:1 earlier:1 cover:1 maximization:2 addressing:1 entry:7 delay:1 successful:1 connect:2 dependency:3 answer:1 combined:1 thanks:1 density:2 sensitivity:8 ie:6 probabilistic:2 infomax:2 michael:1 analogously:1 together:1 concrete:1 again:2 frustration:1 opposed:1 possibly:1 external:1 derivative:5 leading:1 unmix:1 jacobians:2 account:1 potential:15 de:1 speculate:1 coding:2 caused:1 depends:1 blind:2 try:1 view:1 doing:1 tab:7 recover:2 slope:1 contribution:3 square:6 ni:3 il:2 who:1 efficiently:1 yield:2 correspond:1 wal:2 dealt:1 ijkl:1 wab:4 explain:2 datapoint:1 reach:2 synapsis:3 synaptic:10 nonetheless:1 colleague:1 polarizing:1 involved:2 naturally:2 gain:1 recall:1 occured:1 worry:1 feed:1 ta:15 dt:2 follow:2 response:10 synapse:11 evaluated:1 done:2 though:1 refractoriness:1 just:6 biomedical:1 implicit:1 until:2 banerjee:2 somehow:1 riding:1 effect:8 former:1 hence:1 spatially:1 deal:1 during:1 self:1 timingdependent:2 m:3 hong:1 trying:1 ay:1 occuring:1 variational:2 fi:5 spiking:10 camb:1 jl:1 analog:1 resting:1 fk:1 nonlinearity:1 had:2 cuny:1 similarity:1 surface:1 add:1 something:3 multivariate:2 reverse:1 came:1 yi:4 captured:1 additional:2 preceding:1 determine:1 full:4 match:3 post:3 bigger:1 prediction:5 involving:2 variant:1 optimisation:2 confront:1 poisson:1 admittance:1 sometimes:1 represent:1 achieved:1 cell:1 addition:2 interval:2 else:1 singular:1 source:1 pyramidal:1 rka:2 sure:1 induced:1 thing:1 call:1 noting:1 symmetrically:1 feedforward:5 split:1 easy:2 affect:3 xj:7 idea:1 whether:1 crutchfield:1 passed:1 york:2 hessian:2 depression:2 electrotonic:2 ignored:1 clear:2 amount:1 transforms:1 simplest:1 generate:1 inhibitory:1 sign:1 neuroscience:1 delta:1 correctly:1 write:3 key:1 soma:6 threshold:11 falling:1 demultiplexing:5 deneve:1 sum:3 utilising:1 inverse:10 ca2:1 place:1 reader:1 separation:1 draw:1 appendix:3 entirely:1 layer:2 bound:1 followed:1 tackled:1 activity:2 strength:1 rkl:2 eisele:1 kronecker:1 worked:1 segev:1 multiplexing:1 tal:2 interim:1 department:1 developing:1 according:3 conjugate:1 membrane:10 cable:1 modification:3 explained:1 spiked:1 taken:1 demultiplex:3 equation:3 count:2 know:2 end:1 operation:1 robustly:1 original:3 top:2 running:1 tony:1 graphical:1 inspirational:1 giving:2 concatenated:1 bl:3 tensor:1 objective:3 move:1 question:1 quantity:2 spike:84 occurs:2 receptive:1 concentration:1 dependence:4 usual:1 gradient:16 separate:1 unable:1 simulated:2 presynaptic:2 barber:3 cellular:1 argue:1 reason:1 maximising:7 assuming:1 index:6 happy:1 difficult:1 hlog:1 relate:1 rise:1 ba:3 calcium:2 neuron:32 observation:1 fabian:1 looking:1 incorporated:1 jim:1 wkl:19 redwood:1 somatic:1 sharp:3 wac:1 introduced:3 bk:1 pair:1 david:1 kl:9 deletion:1 address:1 able:1 usually:1 pattern:3 below:1 tb:3 memory:5 critical:1 natural:5 lk:8 started:1 extract:1 isn:1 nice:2 understanding:1 l2:1 theis:1 determining:1 camino:1 embedded:3 mixed:3 interesting:1 analogy:1 integrate:2 rni:2 principle:1 gi0:3 share:1 course:1 changed:1 excitatory:1 supported:1 last:3 lef:1 institute:1 explaining:2 taking:1 emerge:1 absolute:1 emphasise:1 distributed:2 slice:1 feedback:2 cortical:1 transition:5 forward:1 made:1 adaptive:1 author:1 invertable:4 far:3 approximate:4 ignore:1 implicitly:2 pseudoinverse:1 assumed:1 ipsp:1 don:2 surya:1 continuous:1 why:2 impedance:1 learn:2 nature:2 robust:1 rearranging:1 ca:1 menlo:1 symmetry:1 ignoring:1 dendrite:1 du:7 gerstner:2 complex:2 froemke:2 anthony:1 cl:3 did:2 spread:1 neurosci:1 neuronal:1 tl:20 ny:1 explicit:1 lie:2 vanish:1 jacobian:6 tin:6 learns:1 bij:1 theorem:1 formula:1 down:1 xt:1 offset:1 organisation:1 deconvolution:1 adding:1 sponses:1 hui:1 disregarded:1 smoothly:1 depicted:1 tc:1 simply:1 explore:2 visual:2 happening:1 expressed:1 partially:1 corresponds:1 goal:1 jeff:1 replace:1 change:9 included:3 determined:2 except:1 specifically:2 called:4 total:2 experimental:3 college:1 rename:2 correlated:1 |
1,844 | 2,675 | Probabilistic computation in spiking populations
Richard S. Zemel
Dept. of Comp. Sci.
Univ. of Toronto
Quentin J. M. Huys
Gatsby CNU
UCL
Rama Natarajan
Dept. of Comp. Sci.
Univ. of Toronto
Peter Dayan
Gatsby CNU
UCL
Abstract
As animals interact with their environments, they must constantly update
estimates about their states. Bayesian models combine prior probabilities, a dynamical model and sensory evidence to update estimates optimally. These models are consistent with the results of many diverse
psychophysical studies. However, little is known about the neural representation and manipulation of such Bayesian information, particularly
in populations of spiking neurons. We consider this issue, suggesting a
model based on standard neural architecture and activations. We illustrate the approach on a simple random walk example, and apply it to
a sensorimotor integration task that provides a particularly compelling
example of dynamic probabilistic computation.
Bayesian models have been used to explain a gamut of experimental results in tasks which
require estimates to be derived from multiple sensory cues. These include a wide range
of psychophysical studies of perception;13 motor action;7 and decision-making.3, 5 Central
to Bayesian inference is that computations are sensitive to uncertainties about afferent and
efferent quantities, arising from ignorance, noise, or inherent ambiguity (e.g., the aperture
problem), and that these uncertainties change over time as information accumulates and
dissipates. Understanding how neurons represent and manipulate uncertain quantities is
therefore key to understanding the neural instantiation of these Bayesian inferences.
Most previous work on representing probabilistic inference in neural populations has focused on the representation of static information.1, 12, 15 These encompass various strategies
for encoding and decoding uncertain quantities, but do not readily generalize to real-world
dynamic information processing tasks, particularly the most interesting cases with stimuli changing over the same timescale as spiking itself.11 Notable exceptions are the recent, seminal, but, as we argue, representationally restricted, models proposed by Gold and
Shadlen,5 Rao,10 and Deneve.4
In this paper, we first show how probabilistic information varying over time can be represented in a spiking population code. Second, we present a method for producing spiking
codes that facilitate further processing of the probabilistic information. Finally, we show
the utility of this method by applying it to a temporal sensorimotor integration task.
1
TRAJECTORY ENCODING AND DECODING
We assume that population spikes R(t) arise stochastically in relation to the trajectory X(t)
of an underlying (but hidden) variable. We use RT and XT for the whole trajectory and
spike trains respectively from time 0 to T . The spikes RT constitute the observations and
are assumed to be probabilistically related to the signal by a tuning function f (X, ? i ):
P (R(i, T )|X(T )) ? f (X, ?i )
(1)
for the spike train of the ith neuron, with parameters ?i . Therefore, via standard Bayesian
inference, RT determines a distribution over the hidden variable at time T , P (X(T )|RT ).
We first consider a version of the dynamics and input coding that permits an analytical
examination of the impact of spikes. Let X(t) follow a stationary Gaussian process such
that the joint distribution P (X(t1 ), X(t2 ), . . . , X(tm )) is Gaussian for any finite collection
of times, with a covariance matrix which depends on time differences: Ctt0 = c(|t ? t0 |).
Function c(|?t|) controls the smoothness of the resulting random walks. Then,
R
P (X(T )|RT ) ? p(X(T )) X(T ) dX(T )P (RT |X(T ))P (X(T )|X(T ))
(2)
where P (X(T )|X(T )) is the distribution over the whole trajectory X(T ) conditional on
the value of X(T ) at its end point. If RT are a set of conditionally independent inhomogeneous Poisson processes, we have
Q
P R
P (RT |X(T )) ? i? f (X(ti? ), ?i ) exp ? i ? d? f (X(? ), ?i ) ,
(3)
where ti? ?? are the spike times ? of neuron i in RT . Let ? = [X(ti? )] be the vector of
stimulus positions at the times at which we observed a spike and ? = [?(ti? )] be the vector
2
2
of spike positions. If the tuning
P Rfunctions are Gaussian f (X, ?i ) ? exp(?(X ? ?i ) /2? )
and sufficiently dense that i ? d? f (X, ?i ) is independent of X (a standard assumption
in population coding), then P (RT |X(T )) ? exp(?k? ? ?k2 /2? 2 ) and in Equation 2, we
can marginalize out X(T ) except at the spike times ti? :
R
2
?1
P (X(T )|RT ) ? p(X(T )) ? d? exp ?[?, X(T )]T C 2 [?, X(T )] ? k???k
(4)
2
2?
and C is the block covariance matrix between X(ti? ), x(T ) at the spike times [tt? ] and the
final time T . This Gaussian integral has P (X(T )|RT ) ? N (?(T ), ?(T )), with
?(T ) = CT t (Ctt + I? 2 )?1 ? = k?
?(T ) = CT T ? kCtT
(5)
CT T is the T, T th element of the covariance matrix and CT t is similarly a row vector. The
dependence in ? on past spike times is specified chiefly by the inverse covariance matrix,
and acts as an effective kernel (k). This kernel is not stationary, since it depends on factors
such as the local density of spiking in the spike train RT .
For example, consider where X(t) evolves according to a diffusion process with drift:
dX = ??Xdt + ? dN (t)
(6)
where ? prevents it from wandering too far, N (t) is white Gaussian noise with mean zero
and ?2 variance. Figure 1A shows sample kernels for this process.
Inspection of Figure 1A reveals some important traits. First, the monotonically decreasing
kernel magnitude as the time span between the spike and the current time T grows matches
the intuition that recent spikes play a more significant role in determining the posterior over
X(T ). Second, the kernel is nearly exponential, with a time constant that depends on the
time constant of the covariance function and the density of the spikes; two settings of these
parameters produced the two groupings of kernels in the figure. Finally, the fully adaptive
kernel k can be locally well approximated by a metronomic kernel k<R> (shown in red in
Figure 1A) that assumes regular spiking. This takes advantage of the general fact, indicated
by the grouping of kernels, that the kernel depends weakly on the actual spike pattern, but
strongly on the average rate. The merits of the metronomic kernel are that it is stationary
and only depends on a single mean rate rather than the full spike train RT . It also justifies
s
Kernels k and k
?0.5
C
5
0
0.03 0.06 0.09
0.04 0.06 0.08
t?t
Time
spike
True stimulus and means
0.1
D
Full kernel
E
Regular, stationary kernel
?0.5
0
?0.5
0.03
0.04
0.05
0.06
0.07
Time
0.08
0.09
0
0.5
Space
0
Space
?4
10
Space
Variance ratio
10
?2
10
0.5
B
?2 / ?2
Kernel size (weight)
A
0.1
0
0.5
0.03
0.04
0.05
0.06
0.07
Time
0.08
0.09
0.1
Figure 1: Exact and approximate spike decoding with the Gaussian process prior. Spikes
are shown in yellow, the true stimulus in green, and P (X(T )|RT ) in gray. Blue: exact
inference with nonstationary and red: approximate inference with regular spiking. A Kernel samples for a diffusion process as defined by equations 5, 6. B, C: Mean and variance
of the inference. D: Exact inference with full kernel k and E: approximation based on
metronomic kernel k<R> . (Equation 7).
the form of decoder used for the network model in the next section.6 Figure 1D shows an
example of how well Equation 5 specifies a distribution over X(t) through very few spikes.
Finally, 1E shows a factorized approximation with the stationary kernel similar to that used
by Hinton and Brown6 and in our recurrent network:
Pt
Q
s
P? (X(t)|R(t)) ? i f (X, ?i ) j=0 kj tij = exp(?E(X(t), R(t), t)),
(7)
By design, the mean is captured very well, but not the variance, which in this example
grows too rapidly for long interspike intervals (Figure 1B, C). Using a slower kernel improves performance on the variance, but at the expense of the mean. We thus turn to the network model with recurrent connections that are available to reinstate the spike-conditional
characteristics of the full kernel.
2
NETWORK MODEL FORMULATION
Above we considered how population spikes RT specify a distribution over X(T ). We now
extend this to consider how interconnected populations of neurons can specify distributions
over time-varying variables. We frame the problem and our approach in terms of a two-level
network, connecting one population of neurons to another; this construction is intended to
apply to any level of processing. The network maps input population spikes R(t) to output
population spikes S(t), where input and output evolve over time. As with the input spikes,
ST indicates the output spike trains from time 0 to T , and these output spikes are assumed
to determine a distribution over a related hidden variable.
For the recurrent and feedforward computation in the network, we start with the deceptively simple goal9 of producing output spikes in such a way that the distribution
Q(X(T )|ST ) they imply over the same hidden variable X(T ) as the input, faithfully
matches P (X(T )|RT ). This might seem a strange goal, since one could surely just listen to the input spikes. However, in order for the output spikes to track the hidden variable,
the dynamics of the interactions between the neurons must explicitly capture the dynamics
of the process X(T ). Once this ?identity mapping? problem has been solved, more general,
complex computations can be performed with ease. We illustrate this on a multisensory
integration task, tracking a hidden variable that depends on multiple sensory cues.
The aim of the recurrent network is to take the spikes R(t) as inputs, and produce output
spikes that capture the probabilistic dynamics. We proceed in two steps. We first consider
the probabilistic decoding process which turns ST into Q(X(t)|ST ). Then we discuss the
recurrent and feedforward processing that produce appropriate ST given this decoder. Note
that this decoding process is not required for the network processing; it instead provides a
computational objective for the spiking dynamics in the system.
We use a simple log-linear decoder based on a spatiotemporal kernel:6
Q(X(T )|ST ) ? exp(?E(X(T ), ST , T )) , where
P PT
E(X, ST , T ) = j ? =0 S(j, T ? ? )?j (X, ? )
(8)
(9)
is an energy function, and the spatiotemporal kernels are assumed separable: ? j (X, ? ) =
gj (X)?(? ). The spatial kernel gj (X) is related to the receptive field f (X, ?j ) of neuron j
and the temporal kernel ?j (X, ? ) to k<RT > .
The dynamics of processing in the network follows a standard recurrent neural architecture
for modeling cortical responses, in the case that network inputs R(t) and outputs S(t) are
spikes. The effect of a spike on other neurons in the network is assumed to have some
simple temporal dynamics, described here again by the temporal kernel ?(? ):
PT
PT
ri (t) = ? =0 R(i, T ? ? )?(? ) sj (t) = ? =0 S(j, T ? ? )?(? )
where T is the extent of the kernel. The response of an output neuron is governed by a
stochastic spiking rule, where the probability that neuron j spikes at time t is given by:
P
P
P (S(j, t) = 1) = ?(uj (t)) = ? ( i wij ri (t) + k vkj sk (t ? 1))
(10)
where ?() is the logistic function, and W and V are the feedforward and recurrent weights.
If ?(? ) = exp(??? ), then uj (T ) = ?(0)(Wj ? R(T ) + Vj ? S(T )) + ?(1)uj (T ? 1); this
corresponds to a discretization of the standard dynamics for the membrane potential of a
du
leaky integrate-and-fire neuron: ? dtj = ??uj +WR+VS, where the leak ? is determined
by the temporal kernel.
The task of the network is to make Q(X(T )|ST ) of Equation 8 match P (X(T )|RT ) coming from one of the two models above (exact dynamic or approximate stationary kernel).
We measure the discrepancy using the Kullback-Leibler (KL) divergence:
P
J = t KL [P (X(T )|RT )||Q(X(T )|ST )]
(11)
and, as a proof of principle in the experiments below, find optimal W and V by
minimizing the KL divergence J using back-propagation through time (BPTT). In order to implement this in the most straightforward way, we convert the stochastic spikingP
rule (EquationP10) to a deterministic rule via the mean-field assumption: S j (t) =
? ( i wij ri (t) + k vkj sk (t ? 1)). The gradients are tedious, but can be neatly expressed in a temporally recursive form. Note that our current focus in the system is on
the representational capability of the system, rather than its learning. Our results establish
that the system can faithfully represent the posterior distribution. We return to the issue of
more plausible learning rules below.
The resulting network can be seen as a dynamic spiking analogue of the recurrent network
scheme of Pouget et al.:9 both methods formulate feedforward and recurrent connections so
that a simple decoding of the output can match optimal but complex decoding applied to the
inputs. A further advantage of the scheme proposed here is that it facilitates downstream
processing of the probabilistic information, as the objective encourages the formation of
distributions at the output that factorize across the units.
3
RELATED MODELS
Ideas about the representation of probabilistic information in spiking neurons are in vogue.
One treatment considers Poisson spiking in populations with regular tuning functions, assuming that stimuli change slowly compared with the inter-spike intervals.8 This leads
to a Kalman filter account with much formal similarity to the models of P (X(T )|R T ).
However, because of the slow timescale, recurrent dynamics can be allowed to settle to an
underlying attractor. In another approach, the spiking activity of either a single neuron 4 or
a pair of neurons5 is considered as reporting (logarithmic) probabilistic information about
an underlying binary hypothesis. A third treatment proposes that a population of neurons
directly represents the (logarithmic) probability over the state of a hidden Markov model. 10
Our method is closely related to the latter two models. Like Deneve?s4 we consider the
transformation of input spikes to output spikes with a fixed assumed decoding scheme so
that the dynamics of an underlying process is captured. Our decoding mechanism produces
something like the predictive coding apparent in Deneve?s scheme, except that here, a neuron may not need to spike not only if it itself has recently spiked and thereby conveyed
the appropriate information; but also if one of its population neighbors has recently spiked.
This is explicitly captured by the recurrent interactions among the population. Our scheme
also resembles Rao?s10 approach in that it involves population codes. Our representational
scheme is more general, however, in that the spatiotemporal decoder defines the relationship between output spikes and Q(X(T )|ST ), whereas his method assumes a direct encoding, with each output neuron?s activity proportional to log Q(X(T )|ST ). Our decoder
can produce such a direct encoding if the spatial and temporal kernels are delta functions,
but other kernels permit coordination amongst the population to take into account temporal
effects, and to produce higher fidelity in the output distribution.
4
EXPERIMENTS
1. Random walk. We describe two experiments. For ease of presentation and comparison, these simulations treat X(t) as a discrete variable, so that the encoding model is a
hidden Markov model (HMM) rather than the Gaussian process defined above. The first
is a random walk, as in Equation 6 and Figure 1, which allows us to make comparisons
with the exact statistics. In a discrete setting, the walk parameters ? and ? determine the
entries in the transition matrix of the corresponding HMM; in a continuous one, the covariance function C of the Gaussian process. Input spikes are generated according to Gaussian
tuning functions (Equation 1). In the recurrent network model, the spatiotemporal kernels are fixed: the spatial kernels are based on the regular locations of the output units j,
gj (X) = |X ? Xj |2 /(1 + |X ? Xj |2 ) and the temporal kernel is ?(? ) = exp(??? ),
where ? = 2 is set to match the walk dynamics. In the following simulations, the network
contained 20 inputs, 60 states, and 20 outputs.
Results on two walk trajectories with different dynamics are shown in Figure 2. The network is trained on walks with parameters (? = 0.2, ? = 2) that force the state to move to
and remain near the center. Figures 2A & B show that in intervals without input spikes, the
inferred mean quickly shifts towards the center and remains there until evidence is received
in the form of input spikes. The feedforward weights (Fig. 2F) show strong connections
between an input unit and its corresponding output, while the learned recurrent weights
(Fig. 2E) reflect the transition probabilities: units coding for extreme values have strong
connections to those nearer the center, and units with preferred values near the center have
strong self-connections. Fig. 2C&D) shows the results of testing this trained network on
walks with different dynamics (? = 0.8, ? = 7). The resulting mismatch between the
mean approximated trajectory (yellow line) and true stimulus (dashed line) (Fig. 2D), and
the variance (Fig. 2H), shows that the learned weights capture the input dynamics.
A
B
E
F
2
3
10
4
2
6
20
8
0
0.5
Variance of Encoded and Decoded Distributions
1
12
0
14
16
?1
18
8
10
12
14
16
18
20
Full Inference
Approximation
120
12
100
?2
0
10
30
12
?0.5
?4
40
14
10
8
6
?1
?6
16
50
18
80
60
40
4
?1.5
?8
?2
20
140
Full Inference
Approximation
14
Variance
10
Output
Output Neurons
Neurons
Input Neurons
2
8
Variance of Encoded and Decoded Distributions
16
1
6
Output Neurons
H
1.5
Variance
4
6
G
4
4
4
D
Recurrent
Weights
Recurrent Weights
Feed?forward Weights
2
2
C
20
60
?2
2
104
6 20 8
10
30
12
40
14
Output Neurons
Output
Neurons
16 50 18
20
60
20
2
0
0
2
4
6
8
10
Time
12
14
16
18
20
0
0
2
4
6
8
10
12
14
16
18
20
Time
Figure 2: Comparison between full inference using hidden Markov model and approximation using network model. Top Row: Full Inference (A,C) and approximation (B,D) results
from two walks. Input spikes (RT ) are shown as green circles; output spikes (ST > .9) as
magenta stars; true stimulus as dashed line; mean inferred trajectory as red line; mean approximated trajectory as yellow line; distributions P (X(t)|RT ) and Q(X(t)|ST ) at each
timestep in gray. Bottom Row: Feedfoward (E); recurrent weights (F); variance of exact
and approximate inference from walks 1 (G) and 2 (H).
2. Sensorimotor task. We next applied our framework to a recent experiment on probabilistic computation during sensorimotor processing.7 Here, human subjects tried to move
a cursor on a display to a target by moving a (hidden) finger. The cursor was shown before
the start of the movement, it was then hidden, except for one point of blurry visual feedback
in the middle of the movement (with variances 0 = ?0 < ?L < ?M < ?? = ?). Unbeknownst to them, on the onset of movement, the cursor was displaced by ?X, drawn from
a prior distribution P (?X). The subjects must estimate ?X in order to compensate for the
displacement and land the cursor on the target. The key result is that subjects learned and
used the prior information P (?X), and indeed integrated it with the visual information
in a way that was appropriately sensitive to the amount of blur (figure 3A). The authors
showed that a simple Bayesian model could account for these data.
We model a population representation of the 2D cursor position X(t) on the screen.
Two spiking input codes?from vision (RT v ) and proprioception (RT p ), present also
in the absence of visual feedback?are mapped into an output code ST representing
P (X(t)|RT v , RT p ). This is a neural instantiation of Bayesian cue combination, and also
an extension of the previous model to the dynamic case.
The first experiment involved a Gaussian prior distribution: P (?X) ? N (1, .5). During
initial experience subjects learn about this prior from trajectories; we determine the parameters of the HMM. We use BPTT to learn feedforward weights for the input spikes from
the different modalities, and recurrent weights between the output units. The input population had 64 units per modality, while the state space and output population each had 100
units. Input spikes were Poisson based on tuning functions centered on a grid within the
2D space; spatiotemporal kernels were based on the (gridded) output units j. The model
was tested in conditions directly matching the experiment, with the cursor and finger moving along a straight line trajectory from the current model?s estimate of the cursor position,
< X(t) >Q(X(t)|ST ) , to the target location. The model captures the main effects of the
experiment (see Figure 3) with respect to visual blur. The prior was ignored when the sensory evidence was precise (?0 ), it dominated on trials without feedback (?? ), and the two
Final Deviation
0
1
2
1
1
1
0
0
0
?1
0
1
2
?1
0
1
?1
0
2
1
2
Imposed displacement (?X)
Final Deviation
Figure 3: (a) Results for a typical subject from the first Ko? rding-Wolpert experiment,7 for
different degrees of blur in the visual feedback (?{0,M,L,?} ). The average lateral displacement of the cursor from the target location at the trial end, as a function of the imposed
displacement of the cursor from the finger location (?X), which is drawn from N (1, .5).
(b) Model results under the same testing conditions. See text for model details.
(a)
(b)
(c)
?2
0
2
Figure 4: (a) Bimodal prior P (?X) ? N (?2, 0.5) for cursor displacement in second
K?ording-Wolpert experiment.7 (b) Results from human subjects. (c) Model results.
factors combined on intermediate degrees of blurriness.
In the second experiment the prior was bimodal (Fig. 4A) and feedback was blurred (? L ).
For this prior, the final cursor location should be based on the more prevalent displacements, so responses based on optimal inference should be non-linear. may modify the
posterior estimate of cursor location. Finally, its 2DIndeed, this is the case (Fig. 4B;C). Intuitively, the blurry visual feedback inadequately defines the true finger position, and thus
the posterior P (X(t)|RT ) is influenced by the learned bimodal prior; the network model
accurately reconstructs this optimal posterior.
Our model generalizes the simple Bayesian account, and suggests new avenues for predictions. The dynamic nature of the system permits modeling the integration of several visual
cues during the trial, as well as differential effects of the timing of visual feedback. The integration of cues in our model also allows it to capture interactions between them. Finally,
its 2D nature allows our system to model other aspects of combining visual and proprioceptive cues, such as their varying and contrasting degrees of certainty across space. 14
5
DISCUSSION
We proposed a spiking population coding scheme for representing and propagating uncertainty which operates at the fine-grained timescale of individual inter-spike intervals. We
motivated the key spatio-temporal spike kernels in the model from analytical results in a
Gaussian process, and suggested two approximations to the exact decoding provided by
these adaptive spatiotemporal kernels. The first is a regular stationary kernel while the
second is a recurrent network model. We showed how gradient descent can set model parameters to match the requirements on the output distribution and capture the dynamics
underlying a hidden variable. This is a dynamic and spiking extension of DPC,15 and a
population extension of Deneve.4 We showed its proficiency by comparison with exact
inference in a random walk, and a neural model that does not use a population code.
The most important direction concerns biologically plausible learning in the full spiking
form of the model. One possibility is to view spikes as a primitive action chosen by a
neuron. In this case, we can use the analog of direct policy methods in partially observable
Markov decision processes,2 with faithful tracking of X(t) leading to reward. It is also
possible that simpler, Hebbian rules will suffice. A second future direction concerns inference of one variable from another using our spiking population code model. This problem
involves marginalizing over intermediate variables, which is difficult in direct representations of distributions over these variables, due to approximating logs of sums with sums of
logs;10 we are investigating how well our scheme can approximate this computation.
We applied the model to a challenging sensorimotor integration task which has been used to
demonstrate Bayesian inference. Since it offers a dynamic account, we can make a number
of predictions about the consequences of variations to the experiment. Most interesting
would be the case in which a bimodal likelihood is combined with a unimodal (or bimodal)
prior (rather than vice-versa), or indeed two instances of visual feedback during the task.
Acknowledgements
We thank Sophie Deneve and Jon Pillow for helpful discussions. RZ & RN funded by NSERC, CIHR
NET program; PD & QH by Gatsby Charitable Fdtn., BIBA consortium, UCL MB/PhD program.
References
[1] Anderson C.H. & Van Essen D.C. (1994). Neurobiological computational systems. In: Computational Intelligence: Imitating Life, Zurada, Marks, Robinson (ed.), IEEE Press, 213-222.
[2] Baxter, J. & Bartlett, P. (2001). Infinite-horizon policy-gradient estimation. JAIR, 319 - 350.
[3] Carpenter, R. H. S. & Williams, M. L. L. (1995). Neural computation of log likelihood in the
control of saccadic eye movements. Nature, 377: 59-62.
[4] Deneve, S. (2004). Bayesian inference in spiking neurons. NIPS-17.
[5] Gold, JI & Shadlen, MN (2001). Neural computations that underlie decisions about sensory
stimuli. Trends in Cognitive Sciences 5:10-16.
[6] Hinton, GE & Brown, AD (2000) Spiking Boltzmann machines. NIPS-12: 122-129.
[7] K?ording, KP & Wolpert, D (2004) Bayesian integration in sensorimotor learning. Nature
427:244-247.
[8] Latham, P., Deneve, S., & Pouget, A., (2004). Optimal computation with attractor networks. J
Physiology, Paris.
[9] Pouget, A., Zhang, K, Deneve, S, & Latham, P. (1998) Statistically efficient estimation using
population codes. NeuralComputation, 10: 373-401.
[10] Rao, R. (2004). Bayesian computation in recurrent neural circuits. Neural Computation, 16(1).
[11] Rieke, F, Warland, D, de Ruyter v. Steveninck, & Bialek, W. (1999). Spikes. MIT Press.
[12] Sahani, M & Dayan, P (2003) Doubly distributional population codes: Simultaneous representation of uncertainty and multiplicity. Neural Computation 15.
[13] Saunders, J.A. & Knill, D.C. (2001). Perception of 3D surface orientation from skew symmetry.
Vision Research, 41 (24) 3163 - 3183.
[14] Van Beers, R. J., Sittig, A.C., & Denier, J.J. (1999). Integration of propriocetive and visual
position-information J Neurophysiol, 81: 1355-1364.
[15] Zemel, R.S., Dayan, P. & Pouget A. (1998). Probabilistic interpretation of population codes.
Neural Computation, 10, 403-430.
| 2675 |@word trial:3 middle:1 version:1 bptt:2 tedious:1 simulation:2 tried:1 covariance:6 thereby:1 initial:1 zurada:1 ording:2 past:1 current:3 discretization:1 activation:1 dx:2 must:3 readily:1 blur:3 interspike:1 motor:1 update:2 v:1 stationary:7 cue:6 intelligence:1 inspection:1 ith:1 feedfoward:1 provides:2 toronto:2 location:6 simpler:1 proficiency:1 zhang:1 dn:1 along:1 direct:4 differential:1 doubly:1 combine:1 rding:1 inter:2 indeed:2 decreasing:1 little:1 actual:1 provided:1 underlying:5 suffice:1 circuit:1 factorized:1 contrasting:1 transformation:1 temporal:9 certainty:1 ti:6 act:1 k2:1 control:2 unit:9 underlie:1 producing:2 t1:1 before:1 local:1 treat:1 modify:1 timing:1 consequence:1 representationally:1 accumulates:1 encoding:5 might:1 resembles:1 suggests:1 challenging:1 biba:1 ease:2 huys:1 range:1 statistically:1 steveninck:1 faithful:1 testing:2 recursive:1 block:1 implement:1 displacement:6 physiology:1 matching:1 regular:6 consortium:1 marginalize:1 applying:1 seminal:1 map:1 deterministic:1 center:4 imposed:2 straightforward:1 primitive:1 williams:1 focused:1 formulate:1 pouget:4 rule:5 deceptively:1 his:1 quentin:1 population:27 rieke:1 variation:1 pt:4 play:1 construction:1 target:4 exact:8 qh:1 hypothesis:1 element:1 trend:1 approximated:3 natarajan:1 particularly:3 distributional:1 observed:1 role:1 bottom:1 solved:1 capture:6 wj:1 movement:4 intuition:1 environment:1 leak:1 pd:1 reward:1 dynamic:23 trained:2 weakly:1 predictive:1 neurophysiol:1 joint:1 various:1 represented:1 finger:4 train:5 univ:2 effective:1 describe:1 kp:1 zemel:2 formation:1 saunders:1 apparent:1 encoded:2 plausible:2 statistic:1 timescale:3 itself:2 final:4 inadequately:1 advantage:2 analytical:2 net:1 ucl:3 interconnected:1 interaction:3 coming:1 cihr:1 mb:1 combining:1 rapidly:1 representational:2 gold:2 requirement:1 produce:5 rama:1 illustrate:2 recurrent:19 propagating:1 received:1 strong:3 involves:2 direction:2 inhomogeneous:1 closely:1 filter:1 stochastic:2 centered:1 human:2 settle:1 require:1 extension:3 sufficiently:1 considered:2 exp:8 mapping:1 estimation:2 coordination:1 sensitive:2 vice:1 faithfully:2 mit:1 gaussian:11 aim:1 rather:4 varying:3 probabilistically:1 derived:1 focus:1 prevalent:1 indicates:1 likelihood:2 helpful:1 inference:18 dayan:3 integrated:1 hidden:12 relation:1 wij:2 issue:2 among:1 fidelity:1 orientation:1 proposes:1 animal:1 spatial:3 integration:8 field:2 once:1 represents:1 nearly:1 jon:1 discrepancy:1 future:1 t2:1 stimulus:8 richard:1 inherent:1 few:1 divergence:2 individual:1 intended:1 fire:1 attractor:2 possibility:1 essen:1 extreme:1 integral:1 experience:1 walk:12 circle:1 uncertain:2 instance:1 modeling:2 compelling:1 rao:3 deviation:2 entry:1 too:2 optimally:1 spatiotemporal:6 combined:2 st:16 density:2 probabilistic:12 decoding:10 connecting:1 quickly:1 again:1 central:1 ambiguity:1 reflect:1 reconstructs:1 slowly:1 stochastically:1 cognitive:1 leading:1 return:1 suggesting:1 potential:1 account:5 de:1 star:1 coding:5 blurred:1 notable:1 explicitly:2 afferent:1 depends:6 onset:1 ad:1 performed:1 view:1 red:3 start:2 capability:1 variance:12 characteristic:1 yellow:3 generalize:1 bayesian:13 accurately:1 produced:1 trajectory:10 comp:2 straight:1 explain:1 simultaneous:1 influenced:1 ed:1 energy:1 sensorimotor:6 involved:1 proof:1 static:1 efferent:1 treatment:2 listen:1 improves:1 back:1 feed:1 higher:1 jair:1 follow:1 specify:2 response:3 formulation:1 strongly:1 anderson:1 just:1 until:1 propagation:1 defines:2 logistic:1 gray:2 indicated:1 grows:2 facilitate:1 effect:4 brown:1 true:5 leibler:1 proprioceptive:1 ignorance:1 conditionally:1 white:1 during:4 self:1 encourages:1 tt:1 demonstrate:1 latham:2 recently:2 spiking:21 ji:1 extend:1 analog:1 interpretation:1 trait:1 significant:1 versa:1 smoothness:1 tuning:5 grid:1 similarly:1 neatly:1 had:2 funded:1 moving:2 similarity:1 surface:1 gj:3 something:1 posterior:5 recent:3 showed:3 ctt:1 manipulation:1 binary:1 life:1 captured:3 seen:1 surely:1 determine:3 monotonically:1 signal:1 dashed:2 multiple:2 encompass:1 full:9 unimodal:1 hebbian:1 match:6 offer:1 long:1 compensate:1 manipulate:1 impact:1 prediction:2 ko:1 vision:2 poisson:3 kernel:38 represent:2 bimodal:5 whereas:1 fine:1 interval:4 modality:2 appropriately:1 subject:6 facilitates:1 proprioception:1 seem:1 nonstationary:1 near:2 feedforward:6 intermediate:2 baxter:1 xj:2 architecture:2 idea:1 tm:1 avenue:1 shift:1 t0:1 motivated:1 utility:1 bartlett:1 peter:1 wandering:1 proceed:1 constitute:1 action:2 ignored:1 tij:1 amount:1 s4:1 locally:1 specifies:1 delta:1 arising:1 track:1 wr:1 per:1 blue:1 diverse:1 discrete:2 key:3 drawn:2 changing:1 diffusion:2 deneve:8 timestep:1 downstream:1 convert:1 sum:2 inverse:1 uncertainty:4 reporting:1 strange:1 decision:3 ct:4 display:1 activity:2 s10:1 ri:3 dominated:1 aspect:1 span:1 separable:1 according:2 vkj:2 combination:1 membrane:1 across:2 remain:1 sittig:1 evolves:1 making:1 biologically:1 intuitively:1 restricted:1 spiked:2 imitating:1 multiplicity:1 unbeknownst:1 equation:7 remains:1 turn:2 discus:1 mechanism:1 skew:1 merit:1 ge:1 end:2 available:1 generalizes:1 permit:3 apply:2 appropriate:2 blurry:2 gridded:1 slower:1 reinstate:1 rz:1 assumes:2 top:1 include:1 xdt:1 warland:1 uj:4 establish:1 approximating:1 psychophysical:2 objective:2 move:2 quantity:3 spike:52 strategy:1 receptive:1 rt:27 dependence:1 saccadic:1 bialek:1 gradient:3 amongst:1 thank:1 mapped:1 sci:2 lateral:1 decoder:5 hmm:3 argue:1 extent:1 considers:1 assuming:1 code:10 cnu:2 kalman:1 relationship:1 ratio:1 minimizing:1 difficult:1 expense:1 design:1 policy:2 boltzmann:1 neuron:25 observation:1 markov:4 displaced:1 finite:1 descent:1 hinton:2 precise:1 frame:1 rn:1 drift:1 inferred:2 pair:1 required:1 specified:1 kl:3 connection:5 paris:1 learned:4 nearer:1 robinson:1 nip:2 suggested:1 dynamical:1 perception:2 pattern:1 below:2 mismatch:1 program:2 green:2 analogue:1 examination:1 dissipates:1 force:1 dpc:1 mn:1 representing:3 scheme:8 imply:1 temporally:1 eye:1 gamut:1 denier:1 kj:1 sahani:1 text:1 prior:12 understanding:2 acknowledgement:1 evolve:1 determining:1 marginalizing:1 fully:1 interesting:2 proportional:1 integrate:1 degree:3 conveyed:1 consistent:1 shadlen:2 beer:1 principle:1 charitable:1 land:1 row:3 formal:1 wide:1 neighbor:1 leaky:1 van:2 feedback:8 cortical:1 world:1 transition:2 pillow:1 sensory:5 forward:1 collection:1 adaptive:2 author:1 far:1 sj:1 approximate:5 observable:1 preferred:1 aperture:1 kullback:1 neurobiological:1 instantiation:2 reveals:1 investigating:1 assumed:5 spatio:1 factorize:1 continuous:1 sk:2 learn:2 nature:4 ruyter:1 symmetry:1 interact:1 du:1 complex:2 vj:1 dense:1 main:1 whole:2 noise:2 arise:1 knill:1 dtj:1 allowed:1 carpenter:1 fig:7 screen:1 gatsby:3 slow:1 position:6 decoded:2 exponential:1 governed:1 third:1 grained:1 magenta:1 xt:1 evidence:3 grouping:2 concern:2 phd:1 magnitude:1 justifies:1 cursor:12 horizon:1 wolpert:3 logarithmic:2 visual:11 prevents:1 expressed:1 contained:1 nserc:1 tracking:2 partially:1 corresponds:1 chiefly:1 constantly:1 determines:1 conditional:2 goal:1 identity:1 presentation:1 towards:1 absence:1 change:2 determined:1 except:3 typical:1 blurriness:1 operates:1 sophie:1 infinite:1 experimental:1 multisensory:1 exception:1 mark:1 latter:1 dept:2 tested:1 |
1,845 | 2,676 | Using the Equivalent Kernel to Understand
Gaussian Process Regression
Peter Sollich
Dept of Mathematics
King?s College London
Strand, London WC2R 2LS, UK
[email protected]
Christopher K. I. Williams
School of Informatics
University of Edinburgh
5 Forrest Hill, Edinburgh EH1 2QL, UK
[email protected]
Abstract
The equivalent kernel [1] is a way of understanding how Gaussian process regression works for large sample sizes based on a continuum limit.
In this paper we show (1) how to approximate the equivalent kernel of the
widely-used squared exponential (or Gaussian) kernel and related kernels, and (2) how analysis using the equivalent kernel helps to understand
the learning curves for Gaussian processes.
Consider the supervised regression problem for a dataset D with entries (xi , yi ) for i =
1, . . . , n. Under Gaussian Process (GP) assumptions the predictive mean at a test point x ?
is given by
f?(x? ) = k> (x? )(K + ? 2 I)?1 y,
(1)
where K denotes the n ? n matrix of covariances between the training points with entries
k(xi , xj ), k(x? ) is the vector of covariances k(xi , x? ), ? 2 is the noise variance on the
observations and y is a n ? 1 vector holding the training targets. See e.g. [2] for further
details.
We can define a vector of functions h(x? ) = (K + ? 2 I)?1 k(x? ) . Thus we have f?(x? ) =
h> (x? )y, making it clear that the mean prediction at a point x? is a linear combination of
the target values y. Gaussian process regression is thus a linear smoother, see [3, section
2.8] for further details. For a fixed test point x? , h(x? ) gives the vector of weights applied
to targets y. Silverman [1] called h> (x? ) the weight function.
Understanding the form of the weight function is made complicated by the matrix inversion
of K + ? 2 I and the fact that K depends on the specific locations of the n datapoints.
Idealizing the situation one can consider the observations to be ?smeared out? in x-space
at some constant density of observations. In this case analytic tools can be brought to bear
on the problem, as shown below. By analogy to kernel smoothing Silverman [1] called the
idealized weight function the equivalent kernel (EK).
The structure of the remainder of the paper is as follows: In section 1 we describe how
to derive the equivalent kernel in Fourier space. Section 2 derives approximations for the
EK for the squared exponential and other kernels. In section 3 we show how use the EK
approach to estimate learning curves for GP regression, and compare GP regression to
kernel regression using the EK.
1
Gaussian Process Regression and the Equivalent Kernel
It is well known (see e.g. [4]) that the posterior mean for GP regression can be obtained as
the function which minimizes the functional
n
1 X
1
(yi ? f (xi ))2 ,
J[f ] = kf k2H + 2
(2)
2
2?n i=1
where kf kH is the RKHS norm corresponding to kernel k. (However, note that the GP
framework gives much more than just this mean prediction, for example the predictive
variance and the marginal likelihood p(y) of the data under the model.)
Let ?(x) = E[y|x] be the target function for our regression problem and write E[(y ?
f (x))2 ] = E[(y ? ?(x))2 ] + (?(x) ? f (x))2 . Using the fact that the first term on the RHS
is independent of f motivates considering a smoothed version of equation 2,
Z
?
1
J? [f ] =
(?(x) ? f (x))2 dx + kf k2H ,
2? 2
2
where ? has dimensions of the number of observations per unit of x-space
(length/area/volume etc. as appropriate). If we consider kernels that are stationary,
k(x, x0 ) = k(x ? x0 ), the natural basis in which to analyse
equation 1 is the Fourier
R
basis of complex sinusoids so that f (x) is represented as f?(s)e2?is?x ds and similarly for
?(x). Thus we obtain
!
Z
1
? ?
|f?(s)|2
2
J? [f ] =
ds,
|f (s) ? ??(s)| +
2
?2
S(s)
R
as kf k2HR =
|f?(s)|2 /S(s)ds where S(s) is the power spectrum of the kernel k,
?2?is?x
S(s) = k(x)e
dx. J? [f ] can be minimized using calculus of variations to ob?(s) = S(s)?(s)/(? 2 /? + S(s)) which is recognized as the convolution f (x? ) =
tain
f
R
h(x? ? x)?(x)dx. Here the Fourier transform of the equivalent kernel h(x) is
?
h(s)
=
1
S(s)
=
.
S(s) + ? 2 /?
1 + ? 2 /(?S(s))
(3)
?
The term ? 2 /? in the first expression for h(s)
corresponds to the power spectrum of a
white noise process, whose delta-function covariance function becomes a constant in the
Fourier domain. This analysis is known as Wiener filtering; see, e.g. [5, ?14-1]. Notice
that as ? ? ?, h(x) tends to the delta function. If the input density is non-uniform the
analysis above should be interpreted as computing the equivalent kernel for np(x) = ?.
This approximation will be valid if the scale of variation of p(x) is larger than the width of
the equivalent kernel.
2
The EK for the Squared Exponential and Related Kernels
For certain kernels/covariance functions the EK h(x) can be computed exactly by Fourier
inversion. Examples include the Ornstein-Uhlenbeck process in D = 1 with covariance
k(x) = e??|x|
(see [5, p. 326]), splines in D = 1 corresponding
to the regularizer
R
R
kP f k2 = (f (m) )2 dx [1, 6], and the regularizer kP f k2 = (?2 f )2 dx in two dimensions, where the EK is given in terms of the Kelvin function kei [7].
We now consider the commonly used squared exponential (SE) kernel k(r) =
exp(?r 2 /2`2 ), where r 2 = ||x?x0 ||2 . (This is sometimes called the Gaussian or radial basis function kernel.) Its Fourier transform is given by S(s) = (2?`2 )D/2 exp(?2? 2 `2 |s|2 ),
where D denotes the dimensionality of x (and s) space.
From equation 3 we obtain
? SE (s) =
h
1
,
1 + b exp(2? 2 `2 |s|2 )
where b = ? 2 /?(2?`2 )D/2 . We are unaware of an exact result in this case, but the following
initial approximation is simple but effective. For large ?, b will be small. Thus for small
? SE ' 1, but for large s it is approximately 0. The change takes
s = |s| we have that h
place around the point sc where b exp(2? 2 `2 s2c ) = 1, i.e. s2c = log(1/b)/2? 2 `2 . As
? SE between 1 and 0 can be expected
exp(2? 2 `2 s2 ) grows quickly with s, the transition of h
to be rapid, and thus be well-approximated by a step function.
Proposition 1 The approximate form of the equivalent kernel for the squared-exponential
kernel in D-dimensions is given by
s D/2
c
hSE (r) =
JD/2 (2?sc r).
r
Proof: hSE (s) is a function of s = |s| only, and for D > 1 the Fourier integral can
be simplified by changing to spherical polar coordinates and integrating out the angular
variables to give
Z ? ?+1
s
? SE (s) ds
hSE (r) = 2?r
J? (2?rs)h
(4)
r
0
Z sc ?+1
s D/2
s
c
' 2?r
J? (2?rs) ds =
JD/2 (2?sc r).
r
r
0
where ? = D/2 ? 1, J? (z) is a Bessel function of the first kind and we have used the
identity z ?+1 J? (z) = (d/dz)[z ?+1 J?+1 (z)].
Note that in D = 1 by computing the Fourier transform of the boxcar function we obtain
hSE (x) = 2sc sinc(2?sc x) where sinc(z) = sin(z)/z. This is consistent with Proposition
1 and J1/2 (z) = (2/?z)1/2 sin(z). The asymptotic form of the EK in D = 2 is shown in
Figure 2(left) below.
Notice that sc scales as (log(?))1/2 so that the width of the EK (which is proportional to
1/sc ) will decay very slowly as ? increases. In contrast for a spline of order m (with power
spectrum ? |s|?2m ) the width of the EK scales as ??1/2m [1].
If instead of RD we consider the input set to P
be the unit circle, a stationary kernel can
be periodized by the construction kp (x, x0 ) = n?Z k(x ? x0 + 2n?). This kernel will
be represented as a Fourier series (rather than with a Fourier transform) because of the
periodicity. In this case the step function in Fourier space approximation would give rise
to a Dirichlet kernel as the EK (see [8, section 4.4.3] for further details on the Dirichlet
kernel).
We now show that the result of Proposition 1 is asymptotically exact for ? ? ?, and calculate the leading corrections for finite ?. The scaling of the width of the EK as 1/s c suggests
writing hSE (r) = (2?sc )D g(2?sc r). Then from equation 4 and using the definition of sc
?+1
Z ?
z
2?sc s
J? (zs/sc )
g(z) =
ds
sc (2?sc )D 0
z
1 + exp[2? 2 `2 (s2 ? s2c )]
Z ?
u ?+1
J? (zu)
=z
du
(5)
2 `2 s2 (u2 ? 1)]
2?z
1
+
exp[2?
0
c
where we have rescaled s = sc u in the second step. The value of sc , and hence ?, now
enters only in the exponential via a = 2? 2 `2 s2c . For a ? ?, the exponential tends to zero
for u < 1 and to infinity for u > 1. The factor 1/[1 + exp(. . .)] is therefore a step function
?(1 ? u) in the limit and Proposition 1 becomes exact, with g? (z) ? lima?? g(z) =
(2?z)?D/2 JD/2 (z). To calculate corrections to this, one uses that for large but finite a the
difference ?(u) = {1 + exp[a(u2 ? 1)]}?1 ? ?(1 ? u) is non-negligible only in a range
of order 1/a around u = 1. The other factors in the integrand of equation 5 can thus be
Taylor-expanded around that point to give
Z ?
?
X
Ik dk u ?+1
J? (zu) , Ik =
g(z) = g? (z) + z
?(u)(u ? 1)k du
k! duk
2?z
0
u=1
k=0
The problem is thus reduced to calculating the integrals Ik . Setting u = 1 + v/a one has
Z ?
Z 0
vk
1
k
?
1
v
dv
+
dv
ak+1 Ik =
2
1 + exp(v 2 /a + 2v)
0
?a 1 + exp(v /a + 2v)
Z a
Z ?
(?1)k+1 v k
vk
=
dv
+
dv
2
1 + exp(v 2 /a + 2v)
0 1 + exp(?v /a + 2v)
0
In the first integral, extending the upper limit to ? gives an error that is exponentially
small in a. Expanding the remaining 1/a-dependence of the integrand one then gets, to
leading order in 1/a, I0 = c0 /a2 , I1 = c1 /a2 while all Ik with k ? 2 are smaller by
at least 1/a2 . The numerical constants are ?c0 = c1 = ? 2 /24. This gives, using that
(d/dz)[z ?+1 J? (z)] = z ? J? (z) + z ?+1 J??1 (z) = (2? + 1)z ? J? (z) ? z ?+1 J?+1 (z):
Proposition 2 The equivalent kernel for the squared-exponential kernel is given for large
? by hSE (r) = (2?sc )D g(2?sc r) with
n
o
1
z
1
g(z) =
JD/2 (z) + 2 (c0 + c1 (D ? 1))JD/2?1 (z) ? c1 zJD/2 (z) +O( 4 )
D
a
a
(2?z) 2
For e.g. D = 1 this becomes g(z) = ? ?1 {sin(z)/z ? ? 2 /(24a2 )[cos(z) + z sin(z)]}. Here
and in general, by comparing the second part of the 1/a2 correction with the leading order
term, one estimates that the correction is of relative size z 2 /a2 . It will therefore provide
a useful improvement as long as z = 2?sc r < a; for larger z the expansion in powers of
1/a becomes a poor approximation because the correction terms (of all orders in 1/a) are
comparable to the leading order.
2.1
Accuracy of the approximation
To evaluate the accuracy of the approximation we can compute the EK numerically as
follows: Consider a dense grid of points in RD with a sampling density ?grid . For making
2
predictions at the grid points we obtain the smoother matrix K(K + ?grid
I)?1 , where1
2
2
?grid = ? ?grid /?, as per equation 1. Each row of this matrix is an approximation to the
EK at the appropriate location, as this is the response to a y vector which is zero at all points
except one. Note that in theory one should use a grid over the whole of RD but in practice
one can obtain an excellent approximation to the EK by only considering a grid around the
point of interest as the EK typically decays with distance. Also, by only considering a finite
grid one can understand how the EK is affected by edge effects.
2
To understand this scaling of ?grid
consider the case where ?grid > ? which means that the
effective variance at each of the ?grid points per unit x-space is larger, but as there are correspondingly
more points this effect cancels out. This can be understood by imagining the situation where there
2
are ?grid /? independent Gaussian observations with variance ?grid
at a single x-point; this would
2
be equivalent to one Gaussian observation with variance ? . In effect the ? observations per unit
x-space have been smoothed out uniformly.
1
0.16
0.35
0.35
Numerical
Proposition 1
Proposition 2
0.3
0.25
0.14
0.12
0.1
0.08
0.2
0.2
0.15
0.15
0.1
0.1
0.05
0.05
0
0
?0.05
?0.05
?0.1
0
5
10
Numerical
Proposition 1
Proposition 2
0.3
0.25
?0.1
0
15
5
10
15
0.06
0.04
0.02
0
?0.02
Numerical
Proposition 1
Sample
?0.04
?0.5
?0.4
?0.3
?0.2
?0.1
0
0.1
0.2
0.3
0.4
0.5
Figure 1: Main figure: plot of the weight function corresponding to ? = 100 training
points/unit length, plus the numerically computed equivalent kernel at x = 0.0 and the sinc
approximation from Proposition 1. Insets: numerically evaluated g(z) together with sinc
and Proposition 2 approximations for ? = 100 (left) and ? = 104 (right).
Figure 1 shows plots of the weight function for ? = 100, the EK computed on the grid as
described above and the analytical sinc approximation. These are computed for parameter
values of `2 = 0.004 and ? 2 = 0.1, with ?grid /? = 5/3. To reduce edge effects, the
interval [?3/2, 3/2] was used for computations, although only the centre of this is shown
in the figure. There is quite good agreement between the numerical computation and the
analytical approximation, although the sidelobes decay more rapidly for the numerically
computed EK. This is not surprising because the absence of a truly hard cutoff in Fourier
space means one should expect less ?ringing? than the analytical approximation predicts.
The figure also shows good agreement between the weight function (based on the finite
sample) and the numerically computed EK. The insets show the approximation of Proposition 2 to g(z) for ? = 100 (a = 5.67, left) and ? = 104 (a = 9.67, right). As expected, the
addition of the 1/a2 -correction gives better agreement with the numerical result for z < a.
Numerical experiments also show that the mean squared error between the numerically
computed EK and the sinc approximation decreases like 1/ log(?). The is larger than the
na??ve estimate (1/a2 )2 ? 1/(log(?))4 based on the first correction term from Proposition
2, because the dominant part of the error comes from the region z > a where the 1/a
expansion breaks down.
2.2
Other kernels
Our analysis is not in fact restricted to the SE kernel. Consider an isotropic kernel, for
which the power spectrum S(s) depends on s = |s| only. Then we can again define from
equation 3 an effective cutoff sc on the range of s in the EK via ? 2 /? = S(sc ), so that
?
h(s)
= [1 + S(sc )/S(s)]?1 . The EK will then have the limiting form given in Proposi?
approaches a step function ?(sc ? s), i.e. if it becomes infinitely ?steep?
tion 1 if h(s)
around the point s = sc for sc ? ?. A quantitative criterion for this is that the slope
? 0 (sc )| should become much larger than 1/sc , the inverse of the range of the step func|h
? 0 (s) = S 0 (s)S(sc )S ?2 (s)[1 + S(sc )/S(s)]?2 , this is equivalent to requiring
tion. Since h
0
that ?sc S (sc )/4S(sc ) ? ?d log S(sc )/d log sc must diverge for sc ? ?. The result
of Proposition 1 therefore applies to any kernel whose power spectrum S(s) decays more
rapidly than any positive power of 1/s.
A trivial example of a kernel obeying this condition would be a superposition of finitely
many SE kernels with different lengthscales `2 ; the asymptotic behaviour of sc is then
governed by the smallest `. A less obvious case is the ?rational quadratic? k(r) =
[1 + (r/l)2 ]?(D+1)/2 which has an exponentially decaying power spectrum S(s) ?
exp(?2?`s). (This relationship is often used in the reverse direction, to obtain the power
spectrum of the Ornstein-Uhlenbeck (OU) kernel exp(?r/`).) Proposition 1 then applies,
with the width of the EK now scaling as 1/sc ? 1/ log(?).
The previous example is a special case of kernels which can be written
as superpositions
R
of SE kernels with a distribution p(`) of lengthscales `, k(r) = exp(?r 2 /2`2 )p(`) d`.
This is in fact the most general representation for an isotropic kernel which defines a valid
covariance function in any dimension D, see [9, ?2.10]. Such a kernel has power spectrum
Z ?
D/2
`D exp(?2? 2 `2 s2 )p(`) d`
S(s) = (2?)
(6)
0
and one easily verifies that the rational quadratic kernel, which has S(s) ? exp(?2?` 0 s),
is obtained for p(`) ? `?D?2 exp(?`20 /2`2 ). More generally, because the exponential
R 1/s
factor in equation 6 acts like a cutoff for ` > 1/s, one estimates S(s) ? 0 `D p(`) d`
for large s. This will decay more strongly than any power of 1/s for s ? ? if p(`) itself
decreases more strongly than any power of ` for ` ? 0. Any such choice of p(`) will
therefore yield a kernel to which Proposition 1 applies.
3
Understanding GP Learning Using the Equivalent Kernel
We now turn to using EK analysis to get a handle on average case learning curves for Gaussian processes. Here the setup is that a function ? is drawn from a Gaussian process, and we
obtain ? noisy observations of ? per unit x-space at random x locations. We are concerned
with the mean squared error (MSE) between the GP prediction f and ?. Averaging over
the noise process, the x-locations of the training data and the prior over ? we obtain the
average MSE as a function of ?. See e.g. [10] and [11] for an overview of earlier work on
GP learning curves.
To understand the asymptotic behaviour of for large ?, we now approximate Rthe true
GP predictions with the EK predictions from noisy data, given by fEK (x) = h(x ?
x0 )y(x0 )dx0 in the continuum limit of ?smoothed out? input locations. We assume as before
that y = target + noise, i.e. y(x) = ?(x) + ?(x) where E[?(x)?(x0 )] = (??2 /?)?(x ? x0 ).
Here ??2 denotes the true noise variance, as opposed to the noise variance assumed in the
EK; the Rscaling ofR ??2 with ? is explained in footnote 1. For a fixed target ?, the MSE
is = ( dx)?1 [?(x) ? fEK (x)]2 dx. Averaging over the noise process ? and target
function ? gives in Fourier space
Z n
o
2 Z
(? 2 /?)S? (s)/S 2 (s) + ??2 /? 2
2
?
? 2 (s) ds = ?
=
S? (s)[1 ? h(s)]
+ (??2 /?)h
ds
?
[1 + ? 2 /(?S(s))]2
(7)
where S? (s) is the power spectrum of the prior over target functions. In the case S(s) =
S? (s) and ? 2 = ??2 where the kernel is exactly matched to theRstructure of the target,
equation 7 gives the Bayes error B and simplifies to B = (? 2 /?) [1 + ? 2 /(?S(s))]?1 ds
(see also [5, eq. 14-16]). Interestingly, this is just the analogue (for a continuous power
spectrum of the kernel rather than a discrete set of eigenvalues) of the lower bound of [10]
?=2
0.5
0.03
0.025
?
0.02
0.015
0.01
?=4
0.1
0.005
0
?0.005
1
0.05
1
0.5
0.5
0
0
?0.5
?0.5
?1
25
?1
50
100
?
250
500
Figure 2: Left: plot of the asymptotic form of the EK (sc /r)J1 (2?sc r) for D = 2 and
? = 1225. Right: log-log plot of against log(?) for the OU and Matern-class processes
(? = 2, 4 respectively). The dashed lines have gradients of ?1/2 and ?3/2 which are the
predicted rates.
on the MSE of standard GP prediction from finite datasets. In experiments this bound
provides a good approximation to the actual average MSE for large dataset size n [11].
This supports our approach of using the EK to understand the learning behaviour of GP
regression.
Treating the denominator in the expression for B again as a hard cutoff at s = sc , which is
justified for large ?, one obtains for an SE target and learner ? ? 2 sc /? ? (log(?))D/2 /?.
To get analogous predictions for the mismatched case, one can write equation 7 as
Z
Z
[1 + ? 2 /(?S(s))] ? ? 2 /(?S(s))
??2
S? (s)
=
ds +
ds.
?
[1 + ? 2 /(?S(s))]2
[S(s)?/? 2 + 1]2
The first integral is smaller than (??2 /? 2 )B and can be neglected as long as B . In the
second integral we can again make the cutoff
R ? approximation?though now with s having
to be above sc ? to get the scaling ? sc sD?1 S? (s) ds. For target functions with a
?
power-law decay S? (s) ? s?? of the power spectrum at large s this predicts ? sD??
c
(log(?))(D??)/2 . So we generically get slow logarithmic learning, consistent with the
observations in [12]. For D = 1 and an OU target (? = 2) we obtain ? (log(?)) ?1/2 , and
for the Matern-class covariance function k(r) = (1 + r/`) exp(?r/`) (which has power
spectrum ? (3/`2 + 4? 2 s2 )?2 , so ? = 4) we get ? (log(?))?3/2 . These predictions
were tested experimentally using a GP learner with SE covariance function (` = 0.1 and
assumed noise level ? 2 = 0.1) against targets from the OU and Matern-class priors (with
` = 0.05) and with noise level ??2 = 0.01, averaging over 100 replications for each value
of ?. To demonstrate the predicted power-law dependence of on log(?), in Figure 2(right)
we make a log-log plot of against log(?). The dashed lines show the gradients of ?1/2
and ?3/2 and we observe good agreement between experimental and theoretical results
for large ?.
3.1
Using the Equivalent Kernel in Kernel Regression
Above we have used the EK to understand how standard GP regression works. One could
alternatively envisage using P
the EK to perform kernel regression, on given finite data sets,
producing a prediction ??1 i h(x? ? xi )yi at x? . Intuitively this seems appealing as a
cheap alternative to full GP regression, particularly for kernels such as the SE where the
EK can be calculated analytically, at least to a good approximation. We now analyze briefly
how such an EK predictor would perform compared to standard GP prediction.
Letting h?iRdenote averaging over noise, training input points and the test point and setting
f? (x? ) = h(x, x? )?(x)dx, the average MSE of the EK predictor is
P
pred = h[?(x) ? (1/?) i h(x, xi )yi ]2 i
R
?2 R
= h[?(x) ? f? (x)]2 + ?? h2 (x, x0 )dx0 i + ?1 h h2 (x, x0 )? 2 (x0 )dx0 i ? ?1 hf?2 (x)i
Z
Z
?2
(? 2 /?)S? (s)/S 2 (s) + ??2 /? 2
h? 2 i
ds
=
ds
+
2
2
2
?
[1 + ? /(?S(s))]
?
[1 + ? /(?S(s))]2
R
R
R
Here we have set h? 2 i = ( dx)?1 ? 2 (x) dx = S? (s) ds for the spatial average of the
squared target amplitude. Taking the matched case, (S? (s) = S(s) and ??2 = ? 2 ) as an
example, the first term (which is the one we get for the prediction from ?smoothed out?
2 D
training inputs, see eq. 7) is of order ? 2 sD
c /?, while the second one is ? h? isc /?. Thus
both terms scale in the same way, but the ratio of the second term to the first is the signalto-noise ratio h? 2 i/? 2 , which in practice is often large. The EK predictor will then perform
significantly worse than standard GP prediction, by a roughly constant factor, and we have
confirmed this prediction numerically. This result is somewhat surprising given the good
agreement between the weight function h(x? ) and the EK that we saw in figure 1, leading
to the conclusion that the detailed structure of the weight function is important for optimal
prediction from finite data sets.
In summary, we have derived accurate approximations for the equivalent kernel (EK) of
GP regression with the widely used squared exponential kernel, and have shown that the
same analysis in fact extends to a whole class of kernels. We have also demonstrated that
EKs provide a simple means of understanding the learning behaviour of GP regression,
even in cases where the learner?s covariance function is not well matched to the structure
of the target function. In future work, it will be interesting to explore in more detail the use
of the EK in kernel smoothing. This is suboptimal compared to standard GP regression as
we saw. However, it does remain feasible even for very large datasets, and may then be
competitive with sparse methods for approximating GP regression. From the theoretical
point of view, the average error of the EK predictor which we calculated may also provide
the basis for useful upper bounds on GP learning curves.
Acknowledgments: This work was supported in part by the IST Programme of the European Community, under the PASCAL Network of Excellence, IST-2002-506778. This
publication only reflects the authors? views.
References
[1] B. W. Silverman. Annals of Statistics, 12:898?916, 1984.
[2] C. K. I. Williams. In M. I. Jordan, editor, Learning in Graphical Models, pages 599?621.
Kluwer Academic, 1998.
[3] T. J. Hastie and R. J. Tibshirani. Generalized Additive Models. Chapman and Hall, 1990.
[4] F. Girosi, M. Jones, and T. Poggio. Neural Computation, 7(2):219?269, 1995.
[5] A. Papoulis. Probability, Random Variables, and Stochastic Processes. McGraw-Hill, New
York, 1991. Third Edition.
[6] C. Thomas-Agnan. Numerical Algorithms, 13:21?32, 1996.
[7] T. Poggio, H. Voorhees, and A. Yuille. Tech. Report AI Memo 833, MIT AI Laboratory, 1985.
[8] B. Sch?olkopf and A. Smola. Learning with Kernels. MIT Press, 2002.
[9] M. L. Stein. Interpolation of Spatial Data. Springer-Verlag, New York, 1999.
[10] M. Opper and F. Vivarelli. In NIPS 11, pages 302?308, 1999.
[11] P. Sollich and A. Halees. Neural Computation, 14:1393?1428, 2002.
[12] P. Sollich. In NIPS 14, pages 519?526, 2002.
| 2676 |@word briefly:1 version:1 inversion:2 norm:1 seems:1 c0:3 calculus:1 r:2 covariance:9 papoulis:1 initial:1 series:1 rkhs:1 interestingly:1 comparing:1 surprising:2 dx:10 must:1 written:1 additive:1 numerical:8 j1:2 girosi:1 analytic:1 cheap:1 plot:5 treating:1 stationary:2 isotropic:2 provides:1 location:5 become:1 ik:5 replication:1 excellence:1 x0:12 expected:2 rapid:1 roughly:1 k2hr:1 spherical:1 actual:1 considering:3 becomes:5 matched:3 kind:1 interpreted:1 minimizes:1 ringing:1 z:1 quantitative:1 act:1 exactly:2 k2:2 uk:4 unit:6 kelvin:1 producing:1 positive:1 negligible:1 understood:1 before:1 tends:2 limit:4 sd:3 ak:1 interpolation:1 approximately:1 plus:1 suggests:1 co:1 range:3 acknowledgment:1 practice:2 silverman:3 area:1 significantly:1 integrating:1 radial:1 get:7 hse:6 writing:1 equivalent:18 demonstrated:1 dz:2 williams:3 l:1 datapoints:1 handle:1 variation:2 coordinate:1 analogous:1 limiting:1 annals:1 target:15 construction:1 lima:1 exact:3 us:1 agreement:5 approximated:1 particularly:1 predicts:2 enters:1 calculate:2 region:1 decrease:2 rescaled:1 sidelobes:1 neglected:1 predictive:2 yuille:1 learner:3 basis:4 easily:1 represented:2 regularizer:2 kcl:1 describe:1 london:2 effective:3 kp:3 sc:44 lengthscales:2 whose:2 quite:1 widely:2 larger:5 statistic:1 gp:21 analyse:1 transform:4 itself:1 noisy:2 envisage:1 eigenvalue:1 analytical:3 remainder:1 rapidly:2 kh:1 olkopf:1 extending:1 help:1 derive:1 ac:2 finitely:1 school:1 eq:2 predicted:2 come:1 direction:1 stochastic:1 wc2r:1 behaviour:4 proposition:17 correction:7 around:5 hall:1 exp:20 k2h:2 continuum:2 a2:8 smallest:1 polar:1 superposition:2 saw:2 tool:1 reflects:1 smeared:1 brought:1 mit:2 gaussian:12 idealizing:1 rather:2 eks:1 publication:1 derived:1 vk:2 improvement:1 likelihood:1 tech:1 contrast:1 i0:1 typically:1 i1:1 pascal:1 smoothing:2 special:1 spatial:2 marginal:1 having:1 sampling:1 chapman:1 jones:1 cancel:1 future:1 minimized:1 np:1 spline:2 report:1 ve:1 interest:1 generically:1 truly:1 isc:1 accurate:1 integral:5 edge:2 poggio:2 taylor:1 circle:1 theoretical:2 earlier:1 entry:2 uniform:1 predictor:4 density:3 informatics:1 diverge:1 together:1 quickly:1 na:1 squared:10 again:3 opposed:1 slowly:1 worse:1 ek:39 leading:5 depends:2 idealized:1 ornstein:2 tion:2 break:1 matern:3 view:2 analyze:1 competitive:1 decaying:1 bayes:1 complicated:1 hf:1 slope:1 accuracy:2 wiener:1 variance:7 yield:1 confirmed:1 footnote:1 ed:1 definition:1 against:3 obvious:1 e2:1 proof:1 rational:2 dataset:2 dimensionality:1 ou:4 amplitude:1 supervised:1 response:1 evaluated:1 though:1 strongly:2 just:2 angular:1 smola:1 d:15 christopher:1 defines:1 grows:1 effect:4 requiring:1 true:2 hence:1 analytically:1 sinusoid:1 laboratory:1 white:1 sin:4 width:5 criterion:1 generalized:1 hill:2 demonstrate:1 functional:1 overview:1 exponentially:2 volume:1 kluwer:1 numerically:7 ai:2 boxcar:1 rd:3 grid:16 mathematics:1 similarly:1 centre:1 etc:1 dominant:1 posterior:1 reverse:1 certain:1 verlag:1 yi:4 ofr:1 somewhat:1 recognized:1 bessel:1 dashed:2 smoother:2 full:1 academic:1 long:2 dept:1 prediction:15 regression:19 denominator:1 kernel:56 uhlenbeck:2 sometimes:1 c1:4 justified:1 addition:1 interval:1 sch:1 jordan:1 concerned:1 xj:1 hastie:1 suboptimal:1 reduce:1 simplifies:1 proposi:1 expression:2 tain:1 peter:2 york:2 useful:2 generally:1 clear:1 se:11 detailed:1 stein:1 reduced:1 notice:2 delta:2 per:5 tibshirani:1 write:2 discrete:1 affected:1 ist:2 drawn:1 changing:1 cutoff:5 asymptotically:1 inverse:1 place:1 extends:1 forrest:1 ob:1 scaling:4 comparable:1 bound:3 quadratic:2 infinity:1 fourier:13 integrand:2 expanded:1 combination:1 poor:1 smaller:2 sollich:4 remain:1 appealing:1 making:2 dv:4 restricted:1 explained:1 intuitively:1 equation:10 turn:1 vivarelli:1 letting:1 observe:1 appropriate:2 alternative:1 jd:5 thomas:1 denotes:3 dirichlet:2 include:1 remaining:1 graphical:1 calculating:1 approximating:1 eh1:1 dependence:2 gradient:2 distance:1 trivial:1 length:2 relationship:1 ratio:2 ql:1 steep:1 setup:1 holding:1 memo:1 rise:1 motivates:1 perform:3 upper:2 observation:9 convolution:1 datasets:2 finite:7 situation:2 smoothed:4 community:1 pred:1 nip:2 below:2 agnan:1 analogue:1 power:18 natural:1 func:1 prior:3 understanding:4 kf:4 asymptotic:4 relative:1 law:2 expect:1 bear:1 interesting:1 filtering:1 proportional:1 analogy:1 h2:2 consistent:2 editor:1 row:1 periodicity:1 summary:1 supported:1 understand:7 mismatched:1 taking:1 correspondingly:1 sparse:1 edinburgh:2 curve:5 dimension:4 calculated:2 valid:2 transition:1 unaware:1 opper:1 author:1 made:1 commonly:1 simplified:1 kei:1 programme:1 approximate:3 obtains:1 rthe:1 mcgraw:1 assumed:2 xi:6 alternatively:1 spectrum:12 continuous:1 duk:1 expanding:1 du:2 expansion:2 excellent:1 complex:1 imagining:1 mse:6 domain:1 european:1 dense:1 main:1 rh:1 s2:5 noise:11 whole:2 edition:1 verifies:1 slow:1 obeying:1 exponential:10 governed:1 third:1 down:1 specific:1 zu:2 inset:2 decay:6 dk:1 sinc:6 voorhees:1 derives:1 signalto:1 logarithmic:1 explore:1 infinitely:1 strand:1 u2:2 halees:1 applies:3 springer:1 corresponds:1 identity:1 king:1 absence:1 feasible:1 change:1 hard:2 experimentally:1 except:1 uniformly:1 averaging:4 called:3 experimental:1 college:1 support:1 dx0:3 s2c:4 evaluate:1 tested:1 |
1,846 | 2,677 | Exponentiated Gradient Algorithms for
Large-margin Structured Classification
Peter L. Bartlett
U.C.Berkeley
Michael Collins
MIT CSAIL
[email protected]
[email protected]
Ben Taskar
Stanford University
David McAllester
TTI at Chicago
[email protected]
[email protected]
Abstract
We consider the problem of structured classification, where the task is
to predict a label y from an input x, and y has meaningful internal structure. Our framework includes supervised training of Markov random
fields and weighted context-free grammars as special cases. We describe
an algorithm that solves the large-margin optimization problem defined
in [12], using an exponential-family (Gibbs distribution) representation
of structured objects. The algorithm is efficient?even in cases where the
number of labels y is exponential in size?provided that certain expectations under Gibbs distributions can be calculated efficiently. The method
for structured labels relies on a more general result, specifically the application of exponentiated gradient updates [7, 8] to quadratic programs.
1
Introduction
Structured classification is the problem of predicting y from x in the case where y has
meaningful internal structure. For example x might be a word string and y a sequence of
part of speech labels, or x might be a Markov random field and y a labeling of x, or x
might be a word string and y a parse of x. In these examples the number of possible labels
y is exponential in the size of x. This paper presents a training algorithm for a general
definition of structured classification covering both Markov random fields and parsing.
We restrict our attention to linear discriminative classification. We assume that pairs hx, yi
can be embedded in a linear feature space ?(x, y), and that a predictive rule is determined
by a direction (weight vector) w in that feature space. In linear discriminative prediction we
select the y that has the greatest value for the inner product h?(x, y), wi. Linear discrimination has been widely studied in the binary and multiclass setting [6, 4]. However, the case
of structured labels has only recently been considered [2, 12, 3, 13]. The structured-label
case takes into account the internal structure of y in the assignment of feature vectors, the
computation of loss, and the definition and use of margins.
We focus on a formulation where each label y is represented as a set of ?parts?, or equivalently, as a bit-vector. Moreover, we assume that the feature vector for y and the loss
for y are both linear in the individual bits of y. This formulation has the advantage that it
naturally covers both simple labeling problems, such as part-of-speech tagging, as well as
more complex problems such as parsing.
We consider the large-margin optimization problem defined in [12] for selecting the classification direction w given a training sample. The starting-point for these methods is a
primal problem that has one constraint for each possible labeling y; or equivalently a dual
problem where each y has an associated dual variable. We give a new training algorithm
that relies on an exponential-family (Gibbs distribution) representation of structured objects. The algorithm is efficient?even in cases where the number of labels y is exponential
in size?provided that certain expectations under Gibbs distributions can be calculated efficiently. The computation of these expectations appears to be a natural computational
problem for structured problems, and has specific polynomial-time dynamic programming
algorithms for some important examples: for example, the clique-tree belief propagation
algorithm can be used in Markov random fields, and the inside-outside algorithm can be
used in the case of weighted context-free grammars.
The optimization method for structured labels relies on a more general result, specifically
the application of exponentiated gradient (EG) updates [7, 8] to quadratic programs (QPs).
We describe a method for solving QPs based on EG updates, and give bounds on its rate of
convergence. The algorithm uses multiplicative updates on dual parameters in the problem.
In addition to their application to the structured-labels task, the EG updates lead to simple
algorithms for optimizing ?conventional? binary or multiclass SVM problems.
Related work [2, 12, 3, 13] consider large-margin methods for Markov random fields and
(weighted) context-free grammars. We consider the optimization problem defined in [12].
[12] use a row-generation approach based on Viterbi decoding combined with an SMO
optimization method. [5] describe exponentiated gradient algorithms for SVMs, but for
binary classification in the ?hard-margin? case, without slack variables. We show that the
EG-QP algorithm converges significantly faster than the rates shown in [5]. Multiplicative
updates for SVMs are also described in [11], but unlike our method, the updates in [11] do
not appear to factor in a way that allows algorithms for MRFs and WCFGs based on Gibbsdistribution representations. Our algorithms are related to those for conditional random
fields (CRFs) [9]. CRFs define a linear model for structured problems, in a similar way
to the models in our work, and also rely on the efficient computation of marginals in the
training phase. Finally, see [1] for a longer version of the current paper, which includes
more complete derivations and proofs.
2
The General Setting
We consider the problem of learning a function f : X ? Y, where X is a set and Y is a
countable set. We assume a loss function L : X ? Y ? Y ? R+ . The function L(x, y, y?)
measures the loss when y is the true label for x, and y? is a predicted label; typically, y? is
the label proposed by some function f (x). In general we will assume that L(x, y, y?) = 0
for y = y?. Given some distribution over examples (X, Y ) in X ? Y, our aim is to find a
function with low expected loss, or risk, EL(X, Y, f (X)).
We consider functions f which take a linear form. First, we assume a fixed function G
which maps an input x to a set of candidates G(x). For all x, we assume that G(x) ? Y,
and that G(x) is finite. A second component to the model is a feature-vector representation
? : X ? Y ? Rd . Given a parameter vector w ? Rd , we consider functions of the form
fw (x) = arg max h?(x, y), wi.
y?G(x)
Given n independent training examples (xi , yi ) with the same distribution as (X, Y ), we
will formalize a large-margin optimization problem that is a generalization of support vector methods for binary classifiers, and is essentially the same as the formulation in [12]. The
optimal parameters are taken to minimize the following regularized empirical risk function:
X
1
2
kwk + C
max (L(xi , yi , y) ? mi,y (w))
y
2
+
i
where mi,y (w) = hw, ?(xi , yi )i ? hw, ?(xi , y)i is the ?margin? on (i, y) and (z)+ =
max{z, 0}. This optimization can be expressed as the primal problem in Figure 1. Following [12], the dual of this problem is also shown in Figure 1. The dual is a quadratic
Primal
problem:
P
2
minw,? 21 kwk + C i i
Subject to the constraints:
?i, ?y ? G(xi ), hw, ?i,y i ? Li,y ? i
?i, i ? 0
Dual problem:
?), where
P max?? F (?
F (?
?) = C i,y ?i,y Li,y ?
P P
1 2
i,y
j,z ?i,y ?j,z h?i,y , ?j,z i
2C
Subject to the constraints:
X
?i,
?i,y = 1 ; ?i, y, ?i,y ? 0
y
P
?
?i,y where w? is the
Relationship between optimal values: w ? = C i,y ?i,y
?
arg min of the primal problem, and ?
? is the arg max of the dual problem.
Figure 1: The primal and dual problems. We use the definitions Li,y = L(xi , yi , y), and ?i,y =
?(xi , yi ) ? ?(xi , y). We assume that for all i, Li,y = 0 for y = yi . The constant C dictates the
relative penalty for values of the slack variables i which are greater than 0.
program F (?
?) in the dual variables ?i,y for all i = 1 . . . n, y ? G(xi ). The dual variables
for each example are constrained to form a probability distribution over Y.
2.1 Models for structured classification
The problems we are interested in concern structured labels, which have a natural decomposition into ?parts?. Formally, we assume some countable set of parts, R. We also assume
a function R which maps each object (x, y) ? X ? Y to a finite subset of R. Thus R(x, y)
is the set of parts belonging to a particular object. In addition we assume a feature-vector
representation ? of parts: this is a function ? : X ? R ? Rd . The feature vector for an
object (x, y) is then a sum of the feature vectors for its parts, and we also assume that the
loss function L(x, y, y?) decomposes into a sum over parts:
X
X
?(x, y) =
?(x, r)
L(x, y, y?) =
l(x, y, r)
r?R(x,y)
r?R(x,?
y)
Here ?(x, r) is a ?local? feature vector for part r paired with input x, and l(x, y, r) is
a ?local? loss for part r when proposed for the pair (x, y). For convenience we define
indicator variables I(x, y, r) which are 1 if r ? R(x, y), 0 otherwise. We also define sets
R(xi ) = ?y?G(xi ) R(xi , y) for all i = 1 . . . n.
Example 1: Markov Random Fields (MRFs) In an MRF the space of labels G(x),
and their underlying structure, can be represented by a graph. The graph G = (V, E) is
a collection of vertices V = {v1 , v2 , . . . vl } and edges E. Each vertex vi ? V has a set
of possible labels, Yi . The set G(x) is then defined as Y1 ? Y2 . . . ? Yl . Each clique in
the graph has a set of possible configurations: for example, if a particular clique contains
vertices {v3 , v5 , v6 }, the set of possible configurations of this clique is Y3 ? Y5 ? Y6 . We
define C to be the set of cliques in the graph, and for any c ? C we define Y(c) to be the set
of possible configurations for that clique. We decompose each y ? G(x) into a set of parts,
by defining R(x, y) = {(c, a) ? R : c ? C, a ? Y(c), (c, a) is consistent with y}. The
feature vector representation ?(x, c, a) for each part can essentially track any characteristics of the assignment a for clique c, together with any features of the input x. A number
of choices for the loss function l(x, y, (c, a)) arePpossible. For example, consider the Hamming loss used in [12], defined as L(x, y, y?) = i Iyi 6=y?i . To achieve this, first assign each
vertex vi to a single one of the cliques in which it appears. Second, define l(x, y, (c, a)) to
be the number of labels in the assignment (c, a) which are both incorrect and correspond
to vertices which have been assigned to the clique c (note that assigning each vertex to a
single clique avoids ?double counting? of label errors).
Example 2: Weighted Context-Free Grammars (WCFGs). In this example x is an
input string, and y is a ?parse tree? for that string, i.e., a left-most derivation for x under
some context-free grammar. The set G(x) is the set of all left-most derivations for x
Inputs: A learning rate ?.
Data structures: A vector ?? of variables, ?i,r , ?i, ?r ? R(xi ).
? = exp(P
?i,r )/Zi where Zi is a normalization term.
Definitions: ?i,y (?)
r?R(xi ,y)
Algorithm:
? Choose initial values ??1 for the ?i,r variables (these values can be arbitrary).
? For t = 1 . . . T + 1:
P
? For i = 1 . . . n, r ? R(xi ), calculate ?ti,r = y ?i,y (??t )I(xi , y, r).
P
P
t
? Set wt = C
i,r?R(xi ,yi ) ?i,r ?
i,r?R(xi ) ?i,r ?i,r
? For i = 1 . . . n, r ? R(xi ),
t+1
t
calculate updates ?i,r
= ?i,r
+ ?C (li,r + hwt , ?i,r i)
Output: Parameter values wT +1
Figure 2: The EG algorithm for structured problems. We use ?i,r = ?(xi , r) and li,r = l(xi , yi , r).
under the grammar. For convenience, we restrict the grammar to be in Chomsky-normal
form, where all rules in the grammar are of the form hA ? B Ci or hA ? ai, where
A, B, C are non-terminal symbols, and a is some terminal symbol. We take a part r to
be a CF-rule-tuple hA ? B C, s, m, ei. Under this representation A spans words s . . . e
inclusive in x; B spans words s . . . m; and C spans words (m + 1) . . . e. The function
R(x, y) maps a derivation y to the set of parts which it includes. In WCFGs ?(x, r) can
be any function mapping a rule production and its position in the sentence x, to a feature
vector. One example of a loss function would be to define l(x, y, r) to be 1 only if r?s
non-terminal A is not seen spanning words s . . . e in the derivation y. This would lead
to L(x, y, y?) tracking the number of ?constituent errors? in y?, where a constituent is a
(non-terminal, start-point, end-point) tuple such as (A, s, e).
3
EG updates for structured objects
We now consider an algorithm for computing ?
? ? = arg max???
F (?
?), where F (?
?) is the
?
dual form of the maximum margin problem, as in Figure 1. In particular, we are interested
in P
the optimal values of the primal form parameters, which are related to ?
? ? by w? =
?
C i,y ?i,y ?i,y . A key problem is that in many of our examples, the number of dual
variables ?i,y precludes dealing with these variables directly. For example, in the MRF
case or the WCFG cases, the set G(x) is exponential in size, and the number of dual
variables ?i,y is therefore also exponential.
We describe an algorithm that is efficient for certain examples of structured objects such
as MRFs or WCFGs. Instead of representing the ?i,y variables explicitly, we will instead
manipulate a vector ?? of variables ?i,r for i = 1 . . . n, r ? R(xi ). Thus we have one of
these ?mini-dual? variables for each part seen in the training data. Each of the variables
?i,r can take any value in the reals. We now define the dual variables ?i,y as a function of
? which takes the form of a Gibbs distribution:
the vector ?,
P
exp( r?R(xi ,y) ?i,r )
?
P
?i,y (?) = P
.
y 0 exp(
r?R(xi ,y 0 ) ?i,r )
Figure 2 shows an algorithm for maximizing F (?
?). The algorithm defines a sequence of
values ??1 , ??2 , . . .. In the next section we prove that the sequence F (?
?(??1 )), F (?
?(??2 )), . . .
converges to max? F (?
?). The algorithm can be implemented efficiently, independently
of the dimensionality of
? , provided that there is an efficient algorithm for computing
P?
?
?
marginal terms ?i,r = i,y ?i,y (?)I(x
i , y, r) for all i = 1 . . . n, r ? R(xi ), and all ?. A
P
P
?
key property is that the primal parameters w = C i,y ?i,y (?)?i,y = C i ?(xi , yi ) ?
C
P
i,y
?
?i,y (?)?(x
i , y) can be expressed in terms of the marginal terms, because:
X
i,y
?
?i,y (?)?(x
i , y) =
X
i,y
?
?i,y (?)
X
?(xi , r) =
r?R(xi ,y)
X
?i,r ?(xi , r)
i,r?R(xi )
P
P
and hence w = C i ?(xi , yi ) ? C i,r?R(xi ) ?i,r ?(xi , r). The ?i,r values can be calculated for MRFs and WCFGs in many cases, using standard algorithms. For example, in
the WCFG case, the inside-outside algorithm can be used, provided that each part r is a
context-free rule production, as described in Example 2 above. In the MRF case, the ? i,r
values can be calculated efficiently if the tree-width of the underlying graph is small.
?
Note that the main storage requirements of the algorithm in Figure 2 concern the vector ?.
This is a vector which has as many components as there are parts in the training set. In
practice, the number of parts in the training data can become extremely large. Fortunately,
an alternative, ?primal form? algorithm is possible. Rather than explicitly storing the ? i,r
variables, we can store a vector zt of the same dimensionality as wt . The ?i,r values can
be computed from zt . More explicitly, the main body of the algorithm in Figure 2 can be
replaced with the following:
? Set z1 to some initial value. For t = 1 . . . T + 1:
? Set wt = 0
t
t
? For i = 1 . . . n: Compute ?ti,r for r? R(xi ), using ?i,r
= ?C((t ? 1)li,r +
hz , ?i,r i);
P
P
t
Set wt = wt + C
r?R(xi ,yi ) ?i,r ?
r?R(xi ) ?i,r ?i,r
? Set zt+1 = zt + wt
1
It can be verified that if ?i, r, ?i,r
= ?Ch?i,r , z1 i, then this alternative algorithm defines
t
the same sequence of (implicit) ?i,r values, and (explicit) wt values, as the original algorithm. In the next section we show that the original algorithm converges for any choice of
1
initial values ??1 , so this restriction on ?i,r
should not be significant.
4
Exponentiated gradient (EG) updates for quadratic programs
We now prove convergence properties of the algorithm in Figure 2. We show that it is
an instantiation of a general algorithm for optimizing quadratic programs (QPs), which
relies on Exponentiated Gradient (EG) updates [7, 8]. In the general problem we assume a
positive semi-definite matrix A ? Rm?m , and a vector b ? Rm , specifying a loss function
Q(?
?) = b0 ?
? + 12 ?
? 0 A?
?. Here ?
? is an m-dimensional vector of reals. We assume
? is
P that ?
formed by the concatenation of n vectors ?
? i ? Rmi for i = 1 . . . n, where i mi = m.
We assume that each ?
? i lies in a simplex of dimension
mi , so that the feasible set is
mi
X
m
?i,j = 1; for all i, j, ?i,j ? 0}.
(1)
? = {?
?:?
? ? R ; for i = 1 . . . n,
j=1
Our aim is to find arg min???
Q(?
?). Figure 3 gives an algorithm?the ?EG-QP?
?
algorithm?for finding the minimum. In the next section we give a proof of its convergence properties.
The EG-QP algorithm can be used to find the minimum of ?F (?
?), and hence the maximum
of the dual objective F (?
?). We justify the algorithm in Figure 2 by showing that it is
equivalent to minimization of ?F (?
?) using the EG-QP algorithm. We give the following
theorem:
P
P P
Theorem 1 Define F (?
?) = C i,y ?i,y Li,y ? 21 C 2 i,y j,z ?i,y ?j,z h?i,y , ?j,z i,
P
and assume as in section 2 that Li,y =
r?R(xi ,y) l(xi , y, r) and ?(xi , y) =
P
1
?
? (? ) . . . ?
? (??T +1 ) defined by the algorithm
r?R(xi ,y) ?(xi , r). Consider the sequence ?
in Figure 2, and the sequence ?
?1 . . . ?
? T +1 defined by the EG-QP algorithm when applied
to Q(?
?) = ?F (?
?). Then under the assumption that ?
? (??1 ) = ?
? 1 , it follows that ?
? (??t ) = ?
?t
for t = 1 . . . (T + 1).
Inputs: A positive semi-definite matrix A, and a vector b, specifying a loss function
Q(?
?) = b ? ?
? + 12 ?
? 0 A?
?. Each vector ?
? is in ?, where ? is defined in Eq. 1.
Algorithm:
? Initialize ?
? 1 to a point in the interior of ?. Choose a learning rate ? > 0.
? For t = 1 . . . T
? Calculate s?t = ?Q(?
?t ) = b + A?
?t .
P t
t+1
t+1
t
? Calculate ?
?
as: ?i, j, ?i,j = ?i,j
exp{??sti,k }
exp{??sti,j }/ k ?i,k
Output: Return ?
? T +1 .
Figure 3: The EG-QP algorithm for quadratic programs.
Proof. We can write F (?)
? = C i,y ?i,y Li,y ? 21 C 2 k i ?(xi , yi ) ? i,y ?i,y ?(xi , y)k2 . It
t
P
(?
? )
follows that ?F
= CLi,y + Ch?(xi , y), wt i = C r?R(xi ,y) li,r + h?i,r , wt i where as
??i,y
P
P
t
?(xi , y)). The rest of the proof proceeds by inbefore wt = C( i ?(xi , yi ) ? i,y ?i,y
duction; due to space constraints we give a sketch of the proof here. The idea is to show that
?
? (??t+1 ) = ?
? t+1 under the inductive hypothesis that ?(
? ??t ) = ?
? t . This follows immediately
t
t+1
?
?
from the definitions of the mappings ?
? (? ) ? ?(
? ? ) and ?
?t ? ?
? t+1 in the two algot
P
?F
(
?
?
)
rithms, together with the identities sti,y = ? ??i,y = ?C r?R(xi ,y) (li,r + h?i,r , wt i)
P
P
P
t+1
t
and ?i,r
? ?i,r
= ?C (li,r + h?i,r , wt i).
4.1
Convergence of the exponentiated gradient QP algorithm
The following theorem shows how the optimization algorithm converges to an optimal solution. The theorem compares the value of the objective function for the algorithm?s vector
?
? t to the value for a comparison vector u ? ?. (For example, consider u as the solution
of the QP.) The convergence result is in terms of several properties of the algorithm and
the comparison vector u. The distance between u and ?
? 1 is measured using the KullbackLiebler (KL) divergence. Recall
that
the
KL-divergence
between two probability vectors
P
u
?, v? is defined as D(?
u, v?) = i ui log uvii . For sequences of probability vectors, u
? ? ?
with u
? = (?
u1 , . . . , u
?n ) and u
?i = (ui,1 , . . . , ui,m
),
we
can
define
a
divergence
as
the
sum
i
? u, v?) = Pn D(?
u
,
v
?
).
Two
other
key
parameters
of KL-divergences: for u
?, v? ? ?, D(?
i
i
i=1
are ?, the largest eigenvalue of the positive semidefinite symmetric matrix A, and
?))i ? min (?Q(?
?))i ? 2 n max |Aij | + max |bi | .
B = max max (?Q(?
???
?
i
i
ij
i
Theorem 2 For all u
? ? ?,
T
X
? u, ?
1
D(?
?1)
e?B ? 1 ? ?B
Q(?
?1 ) ? Q(?
?T +1 )
Q(?
?t ) ? Q(?
u) +
+ 2 2
.
?B
T t=1
?T
? B (1 ? ?(B + ?)e )
T
Choosing ? = 0.4/(B + ?) ensures that
T
? u, ?
1X
D(?
?1)
Q(?
?1 ) ? Q(?
?T +1 )
Q ?
? T +1 ?
Q(?
?t ) ? Q(?
u) + 2.5(B + ?)
+ 1.5
.
T t=1
T
T
The first lemma we require is due to Kivinen and Warmuth [8].
Lemma 1 For any u
? ? ?,
? u, ?
? u, ?
? ?t , ?
?Q(?
?t ) ? ?Q(?
u) ? D(?
? t ) ? D(?
? t+1 ) + D(?
? t+1 )
We focus on the third term. Define ?(i) Q(?
?) as the segment of the gradient vector corresponding
to
the
component
?
?
of
?,
?
and
define the random variable Xi,t , satisfying
i
t
Pr Xi,t = ? ?(i) Q(?
? ) j = ?i,j .
? ?t , ?
Lemma 2 D(?
? t+1 ) =
n
X
h
log E e
i=1
? ?t , ?
Proof. D(?
? t+1 ) =
n X
X
i=1
=
=
j
n X
X
i=1
n
X
t
?ij
log
t
?ij
?(Xi,t ?EXi,t )
log
?
e?B ? 1 ? ?B
B2
t
?ij
log
X
t
?ik
exp(???i,k )
k
X
X
n
var(Xi,t ).
i=1
t+1
?ij
j
i=1
n
X
i
t
?ik
exp ???i,k +
??
? it
!
? ?i
k
+ ??i,j
!
!
n
h
i e?B ? 1 ? ?B X
?(Xi,t ?EXi,t )
=
log E e
?
var(Xi,t ).
B2
i=1
i=1
This last inequality is at the heart of the proof of Bernstein?s inequality; e.g., see [10].
The second part of the proof of the theorem involves bounding this variance in terms of
the loss. The following lemma relies on the fact that this variance is, to first order, the
decrease in the quadratic loss, and that the second order term in the Taylor series expansion
of the loss is small compared to the variance, provided the steps are not too large. The
lemma and its proof require several definitions. For any d, let ? : Rd ? (0, 1)d be the
? i = exp(?i )/ Pd exp(?j ), for ?? ? Rd . We shall work in the
softmax function, ?(?)
j=1
exponential parameter space: let ??t be the exponential parameters at step t, so that the
updates are ??t+1 = ??t ? ??Q(?
?t ), and
?
? t = ?(??it ). Define the
the QP variables satisfy
i
t
random variables Xi,t,??, satisfying Pr Xi,t,?? = ? ?(i) Q(?
? ) j = ?(??i ) j . This takes
the same values as Xi,t , but
is given by a different exponential
parameter
its distribution
(??i instead of ??it ). Define ??t , ??t+1 = a??t + (1 ? a)??t+1 : a ? [0, 1] .
Lemma 3 For some ?? ? [??t , ??t+1 ],
?
n
X
var(Xi,t ) ? ? 2 (B + ?)
n
X
var(Xi,t,??) ? Q(?
?t ) ? Q(?
?t+1 ),
i=1
i=1
but for all ?? ? [??t , ??t+1 ], var(Xi,t,??) ? e?B var(Xi,t ). Hence,
n
X
var(Xi,t ) ?
i=1
1
Q(?
?t ) ? Q(?
?t+1 ) .
?B
? (1 ? ?(B + ?)e )
Thus, for ? < 0.567/(B + ?), Q(?
?t ) is non-increasing in t.
The proof is in [1]. Theorem 2 follows from an easy calculation.
5
Experiments
We compared an online1 version of the Exponentiated Gradient algorithm with the factored
Sequential Minimal Optimization (SMO) algorithm in [12] on a sequence segmentation
task. We selected the first 1000 sentences (12K words) from the CoNLL-2003 named
entity recognition challenge data set for our experiment. The goal is to extract (multiword) entity names of people, organizations, locations and miscellaneous entities. Each
word is labelled by 9 possible tags (beginning of one of the four entity types, continuation
of one of the types, or not-an-entity). We trained a first-order Markov chain over the tags,
In the online algorithm we calculate marginal terms, and updates to the w t parameters, one
training example at a time. As yet we do not have convergence bounds for this method, but we have
found that it works well in practice.
1
14
14
12
12
10
10
8
8
6
SMO
6
SMO
EG(th-2.7)
EG (eta .5)
EG(th-3)
4
EG (eta 1)
(a)
99
92
78
71
64
57
50
43
36
29
22
8
15
99
92
85
78
71
64
57
50
43
36
29
22
8
0
15
0
1
2
1
EG(th-4.5)
2
85
4
(b)
Figure 4: Number of iterations over training set vs. dual objective for the SMO and EG algorithms.
(a) Comparison with different ? values; (b) Comparison with ? = 1 and different initial ? values.
where our cliques are just the nodes for the tag of each word and edges between tags of
consecutive words. The feature vector for each node assignment consists of the word itself,
its capitalization and morphological features, etc., as well as the previous and consecutive
words and their features. Likewise, the feature vector for each edge assignment consists of
the two words and their features as well as surrounding words.
Figure 4 shows the growth of the dual objective function after each pass through the data
for SMO and EG, for several settings of the learning rate ? and the initial setting of the ?
parameters. Note that SMO starts up very quickly but slows down in a suboptimal region,
while EG lags at the start, but overtakes SMO and achieves a larger than 10% increase in
the value of the objective. These preliminary results suggest that a hybrid algorithm could
get the benefits of both, by starting out with several SMO updates and then switching to EG.
The key issue is to switch from the marginal ? representation SMO maintains to the Gibbs ?
representation that EG uses. We can find ? that produces ? by first computing conditional
?probabilities? that correspond to our marginals (e.g. dividing edge marginals by node
marginals in this case) and then letting ??s be the logs of the conditional probabilities.
References
[1] Long version of this paper. Available at http://www.ai.mit.edu/people/mcollins.
[2] Y. Altun, I. Tsochantaridis, and T. Hofmann. Hidden markov support vector machines. In
ICML, 2003.
[3] Michael Collins. Parameter estimation for statistical parsing models: Theory and practice of
distribution-free methods. In Harry Bunt, John Carroll, and Giorgio Satta, editors, New Developments in Parsing Technology. Kluwer, 2004.
[4] K. Crammer and Y. Singer. On the algorithmic implementation of multiclass kernel-based
vector machines. Journal of Machine Learning Research, 2(5):265?292, 2001.
[5] N. Cristianini, C. Campbell, and J. Shawe-Taylor. Multiplicative updatings for support-vector
learning. Technical report, NeuroCOLT2, 1998.
[6] N. Cristianini and J. Shawe-Taylor. An Introduction to Support Vector Machines and Other
Kernel-Based Learning Methods. Cambridge University Press, 2000.
[7] J. Kivinen and M. Warmuth. Exponentiated gradient versus gradient descent for linear predictors. Information and Computation, 132(1):1?63, 1997.
[8] J. Kivinen and M. Warmuth. Relative loss bounds for multidimensional regression problems.
Journal of Machine Learning Research, 45(3):301?329, 2001.
[9] John Lafferty, Andrew McCallum, and Fernando Pereira. Conditional random fields: Probabilistic models for segmenting and labeling sequence data. In Proceedings of ICML-01, 2001.
[10] D. Pollard. Convergence of Stochastic Processes. Springer-Verlag, 1984.
[11] F. Sha, L. Saul, and D. Lee. Multiplicative updates for large margin classifiers. In COLT, 2003.
[12] B. Taskar, C. Guestrin, and D. Koller. Max margin Markov networks. In NIPS, 2003.
[13] I. Tsochantaridis, T. Hofmann, T. Joachims, and Y. Altun. Support vector machine learning for
interdependent and structured output spaces. ICML, 2004 (To appear).
| 2677 |@word version:3 polynomial:1 decomposition:1 initial:5 configuration:3 contains:1 series:1 selecting:1 current:1 assigning:1 yet:1 parsing:4 john:2 chicago:1 hofmann:2 update:15 discrimination:1 v:1 selected:1 warmuth:3 mccallum:1 beginning:1 node:3 location:1 org:1 become:1 ik:2 incorrect:1 prove:2 consists:2 inside:2 tagging:1 expected:1 terminal:4 increasing:1 provided:5 moreover:1 underlying:2 string:4 finding:1 berkeley:2 y3:1 multidimensional:1 ti:2 growth:1 classifier:2 rm:2 k2:1 appear:2 segmenting:1 positive:3 giorgio:1 local:2 switching:1 might:3 studied:1 specifying:2 bi:1 satta:1 practice:3 definite:2 empirical:1 significantly:1 dictate:1 word:14 chomsky:1 suggest:1 get:1 convenience:2 interior:1 altun:2 tsochantaridis:2 storage:1 context:6 risk:2 restriction:1 conventional:1 map:3 equivalent:1 www:1 crfs:2 maximizing:1 attention:1 starting:2 independently:1 immediately:1 factored:1 rule:5 mcollins:2 programming:1 us:2 hypothesis:1 satisfying:2 recognition:1 updating:1 wcfg:2 taskar:2 calculate:5 region:1 ensures:1 morphological:1 decrease:1 pd:1 ui:3 multiword:1 cristianini:2 dynamic:1 trained:1 solving:1 segment:1 predictive:1 exi:2 represented:2 derivation:5 surrounding:1 describe:4 duction:1 labeling:4 outside:2 choosing:1 lag:1 stanford:2 widely:1 larger:1 otherwise:1 precludes:1 grammar:8 itself:1 online:1 sequence:9 advantage:1 eigenvalue:1 product:1 achieve:1 constituent:2 cli:1 convergence:7 double:1 requirement:1 produce:1 converges:4 ben:1 tti:2 object:7 qps:3 andrew:1 stat:1 measured:1 ij:5 b0:1 eq:1 solves:1 dividing:1 implemented:1 c:1 predicted:1 involves:1 direction:2 stochastic:1 mcallester:2 require:2 hx:1 assign:1 generalization:1 btaskar:1 decompose:1 preliminary:1 considered:1 normal:1 exp:9 viterbi:1 predict:1 mapping:2 algorithmic:1 achieves:1 consecutive:2 hwt:1 estimation:1 label:19 largest:1 weighted:4 minimization:1 mit:3 aim:2 rather:1 pn:1 focus:2 joachim:1 mrfs:4 el:1 vl:1 typically:1 hidden:1 koller:1 interested:2 arg:5 classification:8 dual:18 issue:1 colt:1 development:1 constrained:1 special:1 initialize:1 softmax:1 marginal:4 field:8 y6:1 icml:3 simplex:1 report:1 divergence:4 individual:1 replaced:1 phase:1 organization:1 semidefinite:1 primal:8 chain:1 edge:4 tuple:2 minw:1 tree:3 taylor:3 minimal:1 eta:2 cover:1 assignment:5 vertex:6 subset:1 predictor:1 too:1 combined:1 csail:2 probabilistic:1 yl:1 lee:1 decoding:1 michael:2 together:2 quickly:1 choose:2 return:1 li:13 account:1 harry:1 b2:2 includes:3 satisfy:1 explicitly:3 vi:2 multiplicative:4 kwk:2 start:3 maintains:1 minimize:1 formed:1 variance:3 characteristic:1 efficiently:4 likewise:1 correspond:2 algot:1 definition:6 naturally:1 associated:1 proof:10 mi:5 hamming:1 rithms:1 recall:1 dimensionality:2 segmentation:1 formalize:1 campbell:1 appears:2 supervised:1 formulation:3 just:1 implicit:1 sketch:1 parse:2 ei:1 propagation:1 defines:2 name:1 true:1 y2:1 inductive:1 hence:3 assigned:1 symmetric:1 eg:23 width:1 covering:1 complete:1 recently:1 qp:9 kluwer:1 marginals:4 significant:1 cambridge:1 gibbs:6 ai:2 rd:5 shawe:2 longer:1 iyi:1 carroll:1 etc:1 optimizing:2 store:1 certain:3 verlag:1 inequality:2 binary:4 yi:15 seen:2 minimum:2 greater:1 fortunately:1 guestrin:1 v3:1 fernando:1 semi:2 technical:1 faster:1 calculation:1 long:1 manipulate:1 paired:1 prediction:1 mrf:3 regression:1 essentially:2 expectation:3 iteration:1 normalization:1 kernel:2 addition:2 rest:1 unlike:1 capitalization:1 subject:2 hz:1 lafferty:1 counting:1 bernstein:1 easy:1 switch:1 zi:2 restrict:2 suboptimal:1 inner:1 idea:1 multiclass:3 bartlett:2 penalty:1 peter:1 speech:2 pollard:1 svms:2 continuation:1 http:1 track:1 write:1 shall:1 key:4 four:1 verified:1 v1:1 graph:5 sum:3 sti:3 named:1 family:2 conll:1 bit:2 bound:3 quadratic:7 rmi:1 constraint:4 inclusive:1 tag:4 u1:1 min:3 span:3 extremely:1 structured:19 belonging:1 wi:2 online1:1 pr:2 taken:1 heart:1 slack:2 singer:1 letting:1 end:1 available:1 v2:1 alternative:2 original:2 cf:1 objective:5 v5:1 neurocolt2:1 sha:1 gradient:11 distance:1 concatenation:1 entity:5 y5:1 spanning:1 relationship:1 mini:1 equivalently:2 slows:1 implementation:1 countable:2 zt:4 markov:9 finite:2 descent:1 defining:1 y1:1 overtakes:1 arbitrary:1 david:1 pair:2 kl:3 sentence:2 z1:2 smo:10 bunt:1 nip:1 proceeds:1 challenge:1 program:6 max:12 belief:1 greatest:1 natural:2 rely:1 regularized:1 predicting:1 indicator:1 kivinen:3 hybrid:1 representing:1 technology:1 extract:1 interdependent:1 relative:2 embedded:1 loss:16 generation:1 var:7 versus:1 consistent:1 editor:1 storing:1 production:2 row:1 last:1 free:7 aij:1 exponentiated:9 saul:1 benefit:1 calculated:4 dimension:1 avoids:1 collection:1 clique:11 dealing:1 instantiation:1 discriminative:2 xi:62 decomposes:1 expansion:1 complex:1 main:2 bounding:1 body:1 position:1 pereira:1 explicit:1 exponential:10 candidate:1 lie:1 third:1 hw:3 theorem:7 down:1 specific:1 showing:1 symbol:2 svm:1 concern:2 sequential:1 ci:1 margin:11 expressed:2 v6:1 tracking:1 springer:1 ch:2 relies:5 conditional:4 identity:1 goal:1 miscellaneous:1 labelled:1 feasible:1 hard:1 fw:1 specifically:2 determined:1 wt:13 justify:1 lemma:6 pas:1 meaningful:2 select:1 formally:1 internal:3 support:5 people:2 crammer:1 collins:2 |
1,847 | 2,678 | Probabilistic Inference of Alternative Splicing
Events in Microarray Data
Ofer Shai, Brendan J. Frey, and Quaid D. Morris
Dept. of Electrical & Computer Engineering
University of Toronto, Toronto, ON
Qun Pan, Christine Misquitta, and Benjamin J. Blencowe
Banting & Best Dept. of Medical Research
University of Toronto, Toronto, ON
Abstract
Alternative splicing (AS) is an important and frequent step in mammalian
gene expression that allows a single gene to specify multiple products,
and is crucial for the regulation of fundamental biological processes. The
extent of AS regulation, and the mechanisms involved, are not well understood. We have developed a custom DNA microarray platform for
surveying AS levels on a large scale. We present here a generative model
for the AS Array Platform (GenASAP) and demonstrate its utility for
quantifying AS levels in different mouse tissues. Learning is performed
using a variational expectation maximization algorithm, and the parameters are shown to correctly capture expected AS trends. A comparison of
the results obtained with a well-established but low through-put experimental method demonstrate that AS levels obtained from GenASAP are
highly predictive of AS levels in mammalian tissues.
1 Biological diversity through alternative splicing
Current estimates place the number of genes in the human genome at approximately 30,000,
which is a surprisingly small number when one considers that the genome of yeast, a singlecelled organism, has 6,000 genes. The number of genes alone cannot account for the complexity and cell specialization exhibited by higher eukaryotes (i.e. mammals, plants, etc.).
Some of that added complexity can be achieved through the use of alternative splicing,
whereby a single gene can be used to code for a multitude of products.
Genes are segments of the double stranded DNA that contain the information required by
the cell for protein synthesis. That information is coded using an alphabet of 4 (A, C, G,
and T), corresponding to the four nucleotides that make up the DNA. In what is known
as the central dogma of molecular biology, DNA is transcribed to RNA, which in turn is
translated into proteins. Messenger RNA (mRNA) is synthesized in the nucleus of the cell
and carries the genomic information to the ribosome. In eukaryotes, genes are generally
comprised of both exons, which contain the information needed by the cell to synthesize
proteins, and introns, sometimes referred to as spacer DNA, which are spliced out of the
pre-mRNA to create mature mRNA. An estimated 35%-75% of human genes [1] can be
C1
(a)
C1
A
C1
C2
C1 A 3? C2
(b)
(c)
C2
C1 A 3?
A1
A2
C1
C2
C1
A1
C1
A2
C2
C2
C1
(d)
C1 A 5? C2
A 5? C2
C1
C1
C1
C2
C2
A
C2
C2
C2
C1
C2
Figure 1: Four types of AS. Boxes represent exons and lines represent introns, with the possible
splicing alternatives indicated by the connectors. (a) Single cassette exon inclusion/exclusion. C1
and C2 are constitutive exons (exons that are included in all isoforms) and flank a single alternative
exon (A). The alternative exon is included in one isoform and excluded in the other. (b) Alternative
3? (or donor) and alternative 5? (acceptor) splicing sites. Both exons are constitutive, but may contain alternative donor and/or acceptor splicing sites. (c) Mutually exclusive exons. One of the two
alternative exons (A1 and A2 ) may be included in the isoform, but not both. (d) Intron inclusion. An
intron may be included in the mature mRNA strand.
spliced to yield different combinations of exons (called isoforms), a phenomenon referred
to as alternative splicing (AS). There are four major types of AS as shown in Figure 1.
Many multi-exon genes may undergo more than one alternative splicing event, resulting in
many possible isoforms from a single gene. [2]
In addition to adding to the genetic repertoire of an organism by enabling a single gene to
code for more than one protein, AS has been shown to be critical for gene regulation, contributing to tissue specificity, and facilitating evolutionary processes. Despite the evident
importance of AS, its regulation and impact on specific genes remains poorly understood.
The work presented here is concerned with the inference of single cassette exon AS levels
(Figure 1a) based on data obtained from RNA expression arrays, also known as microarrays.
1.1 An exon microarray data set that probes alternative splicing events
Although it is possible to directly analyze the proteins synthesized by a cell, it is easier, and
often more informative, to instead measure the abundance of mRNA present. Traditionally,
gene expression (abundance of mRNA) has been studied using low throughput techniques
(such as RT-PCR or Northern blots), limited to studying a few sequences at a time and
making large scale analysis nearly impossible.
In the early 1990s, microarray technology emerged as a method capable of measuring the
expression of thousands of DNA sequences simultaneously. Sequences of interest are deposited on a substrate the size of a small microscope slide, to form probes. The mRNA
is extracted from the cell and reverse-transcribed back into DNA, which is labelled with
red and green fluorescent dye molecules (cy3 and cy5 respectively). When the sample of
tagged DNA is washed over the slide, complementary strands of DNA from the sample hybridize to the probes on the array forming A-T and C-G pairings. The slide is then scanned
and the fluorescent intensity is measured at each probe. It is generally assumed that the
intensity measure at the probe is linearly related to the abundance of mRNA in the cell over
a wide dynamic range.
Despite significant improvements in microarray technologies in recent years, microarray
data still presents some difficulties in analysis. Low measurements tend to have extremely
low signal to noise ratio (SNR) [7] and probes often bind to sequences that are very similar,
but not identical, to the one for which they were designed (a process referred to as cross-
C1
A
C1
A
C1:A
C1
C2
C2
3 Body probes
A:C2
A
C2
2 Inclusion junction probes
C1:C2
C1
C2
1 Exclusion junction probe
Figure 2: Each alternative splicing event is studied using six probes. Probes were chosen to measure
the expression levels of each of the three exons involved in the event. Additionally, 3 probes are used
that target the junctions that are formed by each of the two isoforms. The inclusion isoform would
express the junctions formed by C1 and A, and A and C2 , while the exclusion isoform would express
the junction formed by C1 and C2
hybridization). Additionally, probes exhibit somewhat varying hybridization efficiency,
and sequences exhibit varying labelling efficiency.
To design our data sets, we mined public sequence databases and identified exons that were
strong candidates for exhibiting AS (the details of that analysis are provided elsewhere
[4, 3]). Of the candidates, 3,126 potential AS events in 2,647 unique mouse genes were
selected for the design of Agilent Custom Oligonucleotide microarray. The arrays were
hybridized with unamplified mRNA samples extracted from 10 wild-type mouse tissues
(brain, heart, intestine, kidney, liver, lung, salivary gland, skeletal muscle, spleen, and
testis). Each AS event has six target probes on the arrays, chosen from regions of the
C1 exon, C2 exon, A exon, C1 :A splice junction, A:C2 splice junction, and C1 :C2 splice
junction, as shown in Figure 2.
2 Unsupervised discovery of alternative splicing
With the exception of the probe measuring the alternative exon, A (Figure 2), all probes
measure sequences that occur in both isoforms. For example, while the sequence of the
probe measuring the junction A:C1 is designed to measure the inclusion isoform, half of it
corresponds to a sequence that is found in the exclusion isoform. We can therefore safely
assume that the measured intensity at each probe is a result of a certain amount of both
isoforms binding to the probe. Due to the generally assumed linear relationship between
the abundance of mRNA hybridized at a probe and the fluorescent intensity measured,
we model the measured intensity as a weighted sum of the overall abundance of the two
isoforms.
A stronger assumption is that of a single, consistent hybridization profile for both isoforms
across all probes and all slides. Ideally, one would prefer to estimate an individual hybridization profile for each AS event studied across all slides. However, in our current
setup, the number of tissues is small (10), resulting in two difficulties. First, the number of
parameters is very large when compared to the number of data point using this model, and
second, a portion of the events do not exhibit tissue specific alternative splicing within our
small set of tissues. While the first hurdle could be accounted for using Baysian parameter
estimation, the second cannot.
2.1 GenASAP - a generative model for alternative splicing array platform
Using the setup described above, the expression vector x, containing the six microarray
measurements as real numbers, can be decomposed as a linear combination of the abundance of the two splice isoforms, represented by the real vector s, with some added noise:
x = ?s + noise, where ? is a 6 ? 2 weight matrix containing the hybridization profiles for
s1
x^ C
x^ C
1
2
s2
x^ A
x^ C :A
x^ A:C
2
x^ C1:C2
2
xC1:C2
1
r
xC
xC
2
xA
xC :A
xA:C
oC1
oC2
oA
oC :A
oA:C
1
1
1
2
oC :C
1
2
?n 2
Figure 3: Graphical model for alternative splicing. Each measurement in the observed expression
profile, x, is generated by either using a scale factor, r, on a linear combination of the isoforms, s, or
drawing randomly from an outlier model. For a detailed description of the model, see text.
the two isoforms across the six probes. Note that we may not have a negative amount of
a given isoform, nor can the presence of an isoform deduct from the measured expression,
and so both s and ? are constrained to be positive.
Expression levels measured by microarrays have previously been modelled as having
expression-dependent noise [7]. To address this, we rewrite the above formulation as
x = r(?s + ?),
(1)
where r is a scale factor and ? is a zero-mean normally distributed random variable with
a diagonal covariance matrix, ?, denoted as p(?) = N (?; 0, ?). The prior distribution for
the abundance of the splice isoforms is given by a truncated normal distribution, denoted
as p(s) ? N (s, 0, I)[s ? 0], where [?] is an indicator function such that [s ? 0] = 1 if
?i, si ? 0, and [s ? 0] = 0 otherwise.
Lastly, there is a need to account for aberrant observations (e.g. due to faulty probes, flakes
of dust, etc.) with an outlier model. The complete GenASAP model (shown in Figure 3)
accounts for the observations as the outcome of either applying equation (1) or an outlier
model. To avoid degenerate cases and ensure meaningful and interpretable results, the
number of faulty probes considered for each AS event may not exceed two, as indicated by
the filled-in square constraint node in Figure 3.
The distribution of x conditional on the latent variables, s, r, and o, is:
p(x|s, r, o) =
Y
N (xi ; r?i s, r2 ?i )[oi =0] N (xi ; Ei , Vi )[oi =1] ,
(2)
i
where oi ? {0, 1} is a bernoulli random variable indicating if the measurement at probe xi
is the result of the AS model or the outlier model parameterized by p(oi = 1) = ?i . The
parameters of the outlier model, E and V, are not optimized and are set to the mean and
variance of the data.
2.2 Variational learning in the GenASAP model
To infer the posterior distribution over the splice isoform abundances while at the same time
learning the model parameters we use a variational expectation-maximization algorithm
(EM). EM maximizes the log likelihood of the data by iteratively estimating the posterior
distribution of the model given the data in the expectation (E) step, and maximizing the
log likelihood with respect to the parameters, while keeping the posterior fixed, in the
maximization (M) step. Variational EM is used when, as in the case of GenASAP, the exact
posterior is intractable. Variational EM minimizes the free energy of the model, defined as
the KL-divergence between the joint distribution of the latent and observed variables and
the approximation to the posterior under the model parameters [5, 6].
We approximate the true posterior using the Q distribution given by
Q({s(t) }, {o(t) }, {r(t) }) =
T
Y
Q(r(t) )Q(o(t) |r(t) )
t=1
Y
(t)
(t)
Q(si |oi , r(t) )
i
T
?1 Y
=Z (t)
(3)
d
d
(t)
(t)
?(t) ? (t) N (s(t) ; ?(t)
? 0],
ro , ?ro )[s
t=1
where Z is a normalization constant, the superscript d indicates that ? is constrained to be
diagonal, and there are T iid AS events. For computational efficiency, r is selected from
a finite set, r ? {r1 , r2 , . . . , rC } with uniform probability. The variational free energy is
given by
XXZ
Q({s(t) }, {o(t) }, {r(t) })
.
F(Q, P ) =
Q({s(t) }, {o(t) }, {r(t) }) log
P ({s(t) }, {o(t) }, {r(t) }, {x(t) })
s
r
o
(4)
Variational EM minimizes the free energy by iteratively updating the Q distribution?s vari(t)d
(t)d
ational parameters (?(t) , ? (t) , ?ro , and ?ro ) in the E-step, and the model parameters (?,
?, {r1 , r2 , . . . , rC }, and ?) in the M-step. The resulting updates are too long to be shown
in the context of this paper and are discussed in detail elsewhere [3]. A few particular points
regarding the E-step are worth covering in detail here.
If the prior on s was a full normal distribution, there would be no need for a variational
approach, and exact EM is possible. For a truncated normal distribution, however, the mixing proportions, Q(r)Q(o|r) cannot be calculated analytically except for the case where s
is scalar, necessitating the diagonality constraint. Note that if ? was allowed to be a full
covariance matrix, equation (3) would be the true posterior, and we could find the sufficient
statistics of Q(s(t) |o(t) , r(t) ):
?1
T
(t) T ?1
?(t)
(I ? O(t) )?)?1 ?T (I ? O(t) )T ??1 x(t) r(t)
ro = (I + ? (I ? O ) ?
?1
?(t)
ro
= (I + ?T (I ? O(t) )T ??1 (I ? O(t) )?)
(5)
(6)
where O is a diagonal matrix with elements Oi,i = oi . Furthermore, it can be easily
shown that the optimal settings for ?d and ?d approximating a normal distribution with
full covariance ? and mean ? is
?doptimal = ?
?1
?doptimal
= diag(?
(7)
?1
)
(8)
In the truncated case, equation (8) is still true. Equation (7) does not hold, though, and
?doptimal cannot be found analytically. In our experiments, we found that using equation
(7) still decreases the free energy every E-step, and it is significantly more efficient than
using, for example, a gradient decent method to compute the optimal ?d .
Intuitive Weigh Matrix
Optimal Weight Matrix
50
50
40
40
30
30
20
20
10
0
10
Inclusion Isoform
0
Exclusion Isoform
Inclusion Isoform
(a)
Exclusion Isoform
(b)
Figure 4: (a) An intuitive set of weights. Based on the biological background, one would expect to
see the inclusion isoform hybridize to the probes measuring C1 , C2 , A, C1 :A, and A:C2 , while the
exclusion isoform hybridizes to C1 , C2 , and C1 :C2 . (b) The learned set of weights closely agrees
with the intuition, and captures cross hybridization between the probes
Contribution of
exclusion isoform
Contribution of
inclusion isoform
AS model
Original Data
(a)
(b)
RT?PCR
AS model
measurement
prediction
(% exclusion) (% exclusion)
14%
27%
72%
70%
8%
22%
outliers
(c)
Figure 5: Three examples of data cases and their predictions. (a) The data does not follow our notion
of single cassette exon AS, but the AS level is predicted accurately by the model.(b) The probe C1 :A
is marked as outlier, allowing the model to predict the other probes accurately. (c) Two probes are
marked as outliers, and the model is still successful in predicting the AS levels.
3 Making biological predictions about alternative splicing
The results presented in this paper were obtained using two stages of learning. In the first
step, the weight matrix, ?, is learned on a subset of the data that is selected for quality.
Two selection criteria were used: (a) sequencing data was used to select those cases for
which, with high confidence, no other AS event is present (Figure 1) and (b) probe sets
were selected for high expression, as determined by a set of negative controls. The second
selection criterion is motivated by the common assumption that low intensity measurements
are of lesser quality (see Section 1.1). In the second step, ? is kept fixed, and we introduce
the additional constraint that the noise is isotropic (? = ?I) and learn on the entire data
set. The constraint on the noise is introduced to prevent the model from using only a subset
of the six probes for making the final set of predictions.
We show a typical learned set of weights in Figure 4. The weights fit well with our intuition
of what they should be to capture the presence of the two isoforms. Moreover, the learned
weights account for the specific trends in the data. Examples of model prediction based on
the microarray data are shown in Figure 5.
Due to the nature of the microarray data, we do not expect all the inferred abundances to be
equally good, and we devised a scoring criterion that ranks each AS event based on its fit to
the model. Intuitively, given two input vectors that are equivalent up to a scale factor, with
inferred MAP estimations that are equal up to the same scale factor,
P we would like their
scores to be identical. The scoring criterion used, therefore is k (xk ? r?k s)2 /(xk +
Rank
500
1000
2000
5000
10000
15000
20000
30000
Pearson?s correlation
coefficient
0.94
0.95
0.95
0.79
0.79
0.78
0.75
0.65
False positive
rate
0.11
0.08
0.05
0.2
0.25
0.29
0.32
0.42
Table 1: Model performance evaluated at various ranks. Using 180 RT-PCR measurements, we are
able to predict the model?s performance at various ranks. Two evaluation criteria are used: Pearson?s
correlation coefficient between the model?s predictions and the RT-PCR measurements and false
positive rate, where a prediction is considered to be false positive if it is more than 15% away from
the RT-PCR measurement.
r?k s)2 , where the MAP estimations for r and s are used. This scoring criterion can be
viewed as proportional to the sum of noise to signal ratios, as estimated using the two
values given by the observation and the model?s best prediction of that observation.
Since it is the relative amount of the isoforms that is of most interest, we need to use the
inferred distribution of the isoform abundances to obtain an estimate for the relative levels
of AS. It is not immediately clear how this should be done. We do, however, have RTPCR measurements for 180 AS events to guide us (see figure 6 for details). Using the
top 50 ranked RT-PCR measurement, we fit three parameters, {a1 , a2 , a3 }, such that the
s2
+ a3 , where s1 is the
proportion of excluded isoform present, p, is given by p = a1 s1 +a
2 s2
MAP estimation of the abundance of the inclusion isoform, s2 is the MAP estimation of the
abundance of the exclusion isoform, and the RT-PCR measurement are used for target p.
The parameters are fitted using gradient descent on a least squared error (LSE) evaluation
criterion.
We used two criteria to evaluate the quality of the AS model predictions. Pearson?s correlation coefficient (PCC) is used to evaluate the overall ability of the model to correctly
estimate trends in the data. PCC is invariant to affine transformation and so is independent
of the transformation parameters a1 and a3 discussed above, while the parameter a2 was
found to effect PCC very little. The PCC stays above 0.75 for the top two thirds ranked predictions. The second evaluation criterion used is the false positive rate, where a prediction
is considered to be false positive if it is more than 15% away from the RT-PCR measurement. This allows us to say, for example, that if a prediction is within the top 10000, we
are 75% confident that it is within 15% of the actual levels of AS.
4 Summary
We designed a novel AS model for the inference of the relative abundance of two alternatively spliced isoforms from six measurements. Unsupervised learning in the model is
performed using a structured variational EM algorithm, which correctly captures the underlying structure of the data, as suggested by its biological nature. The AS model, though
presented here for a cassette exon AS events, can be used to learn any type of AS, and with
a simple adjustment, multiple types.
The predictions obtained from the AS model are currently being used to verify various
claims about the role of AS in evolution and functional genomics, and to help identify
sequences that affect the regulation of AS.
% Exclusion isoform
RT?PCR measurement Vs. AS model predictions
RT?PCR measurements:
90
80
AS model prediction
Int
es
Te tine
sti
Kid s
n
Sa ey
liva
Br ry
ain
Sp
le
Liv en
er
Mu
sc
Lu le
ng
100
70
60
50
40
30
14 22 27 32 47 46 66 78 63
20
AS model prediction:
10
27 24 26 26 51 75 60 85 100
(a)
0
0
20
40
60
80
RT?PCR measurement
100
(b)
Figure 6: (a) Sample RT-PCR. RNA extracted from the cell is reverse-transcribed to DNA, amplified
and labelled with radioactive or fluorescent molecules. The sample is pulled through a viscous gel in
an electric field (DNA, being an acid, is positively charged). Shorter strands travel further through
the gel than longer ones, resulting in two distinct bands, corresponding to the two isoforms, when
exposed to a photosensitive or x-ray film. (b) A scatter plot showing the RT-PCR measurements as
compared to the AS model predictions. The plot shows all available RT-PCR measurements with a
rank of 8000 or better.
The AS model presented assumes a single weight matrix for all data cases. This is an
oversimplified view of the data, and current work is being carried out in identifying probe
specific expression profiles. However, due to the low dimensionality of the problem (10 tissues, six probes per event), care must be taken to avoid overfitting and to ensure meaningful
interpretations.
Acknowledgments
We would like to thank Wen Zhang, Naveed Mohammad, and Timothy Hughes for their
contributions in generating the data set. This work was funded in part by an operating and
infrastructure grants from the CIHR and CFI, and a operating grants from NSERC and a
Premier?s Research Excellence Award.
References
[1] J. M. Johnson et al. Genome-wide survey of human alternative pre-mrna splicing with exon
junction microarrays. Science, 302:2141?44, 2003.
[2] L. Cartegni et al. Listening to silence and understanding nonsense: exonic mutations that affect
splicing. Nature Gen. Rev., 3:285?98, 2002.
[3] Q. Pan et al. Revealing global regulatory features of mammalian alternative splicing using a
quantitative microarray platform. Molecular Cell, 16(6):929?41, 2004.
[4] Q. Pan et al. Alternative splicing of conserved exons is frequently species specific in human and
mouse. Trends Gen., In Press, 2005.
[5] M. I. Jordan, Z. Ghahramani, T. Jaakkola, and Lawrence K. Saul. An introduction to variational
methods for graphical models. Machine Learning, 37(2):183? 233, 1999.
[6] R. M. Neal and G. E. Hinton. A view of the em algorithm that justifies incremental, sparse, and
other variants. In Learning in Graphical Models. Cambridge, MIT Press, 1998.
[7] D. M. Rocke and B. Durbin. A model for measurement error for gene expression arrays. Journal
of Computational Biology, 8(6):557?69, 2001.
| 2678 |@word pcc:4 stronger:1 proportion:2 covariance:3 mammal:1 carry:1 score:1 genetic:1 tine:1 current:3 aberrant:1 si:2 scatter:1 liva:1 must:1 deposited:1 informative:1 designed:3 interpretable:1 update:1 plot:2 v:1 alone:1 generative:2 selected:4 half:1 xxz:1 isotropic:1 xk:2 eukaryote:2 infrastructure:1 node:1 toronto:4 zhang:1 rc:2 c2:32 pairing:1 wild:1 ray:1 introduce:1 excellence:1 expected:1 nor:1 frequently:1 multi:1 brain:1 ry:1 oversimplified:1 decomposed:1 little:1 actual:1 provided:1 estimating:1 moreover:1 underlying:1 maximizes:1 what:2 viscous:1 surveying:1 isoform:22 minimizes:2 developed:1 transformation:2 safely:1 quantitative:1 every:1 ro:6 control:1 normally:1 medical:1 grant:2 positive:6 understood:2 bind:1 engineering:1 frey:1 despite:2 approximately:1 studied:3 limited:1 range:1 unique:1 acknowledgment:1 hughes:1 cfi:1 significantly:1 acceptor:2 revealing:1 pre:2 confidence:1 specificity:1 protein:5 cannot:4 selection:2 put:1 faulty:2 impossible:1 applying:1 context:1 equivalent:1 map:4 charged:1 maximizing:1 mrna:11 kidney:1 survey:1 identifying:1 immediately:1 array:7 notion:1 traditionally:1 target:3 exact:2 substrate:1 trend:4 synthesize:1 element:1 updating:1 mammalian:3 donor:2 database:1 observed:2 role:1 electrical:1 capture:4 thousand:1 region:1 spliced:3 decrease:1 weigh:1 benjamin:1 intuition:2 mu:1 complexity:2 ideally:1 dynamic:1 dogma:1 segment:1 rewrite:1 exposed:1 predictive:1 efficiency:3 translated:1 exon:24 easily:1 connector:1 joint:1 represented:1 various:3 alphabet:1 distinct:1 sc:1 outcome:1 pearson:3 banting:1 emerged:1 film:1 constitutive:2 drawing:1 otherwise:1 say:1 ability:1 statistic:1 superscript:1 final:1 sequence:10 cihr:1 product:2 frequent:1 gen:2 mixing:1 poorly:1 degenerate:1 amplified:1 description:1 intuitive:2 double:1 r1:2 generating:1 incremental:1 help:1 liver:1 measured:6 sa:1 strong:1 predicted:1 exhibiting:1 closely:1 human:4 public:1 repertoire:1 biological:5 hold:1 considered:3 normal:4 lawrence:1 predict:2 claim:1 major:1 early:1 a2:5 estimation:5 travel:1 currently:1 ain:1 agrees:1 create:1 weighted:1 mit:1 genomic:1 rna:4 avoid:2 varying:2 jaakkola:1 kid:1 improvement:1 bernoulli:1 likelihood:2 indicates:1 sequencing:1 rank:5 brendan:1 inference:3 dependent:1 entire:1 overall:2 denoted:2 platform:4 constrained:2 equal:1 field:1 having:1 ng:1 biology:2 identical:2 unsupervised:2 throughput:1 nearly:1 few:2 wen:1 randomly:1 simultaneously:1 divergence:1 individual:1 quaid:1 interest:2 highly:1 custom:2 evaluation:3 capable:1 nucleotide:1 shorter:1 filled:1 fitted:1 measuring:4 maximization:3 subset:2 hybridizes:1 snr:1 uniform:1 comprised:1 successful:1 johnson:1 too:1 confident:1 fundamental:1 stay:1 probabilistic:1 synthesis:1 mouse:4 squared:1 central:1 containing:2 transcribed:3 account:4 potential:1 diversity:1 coefficient:3 int:1 vi:1 performed:2 view:2 analyze:1 red:1 portion:1 lung:1 shai:1 mutation:1 contribution:3 formed:3 square:1 oi:7 isoforms:16 acid:1 variance:1 yield:1 identify:1 modelled:1 radioactive:1 accurately:2 iid:1 lu:1 worth:1 tissue:8 premier:1 messenger:1 energy:4 involved:2 hybridize:2 exonic:1 dimensionality:1 back:1 higher:1 agilent:1 follow:1 specify:1 formulation:1 evaluated:1 box:1 though:2 done:1 furthermore:1 xa:2 stage:1 lastly:1 correlation:3 ei:1 gland:1 quality:3 indicated:2 yeast:1 effect:1 contain:3 true:3 verify:1 evolution:1 tagged:1 analytically:2 excluded:2 iteratively:2 neal:1 ribosome:1 covering:1 whereby:1 oc:2 criterion:9 evident:1 complete:1 demonstrate:2 necessitating:1 mohammad:1 christine:1 lse:1 variational:10 novel:1 common:1 functional:1 discussed:2 organism:2 interpretation:1 synthesized:2 significant:1 measurement:20 cambridge:1 inclusion:10 funded:1 longer:1 operating:2 etc:2 dust:1 posterior:7 exclusion:12 dye:1 recent:1 reverse:2 certain:1 muscle:1 scoring:3 conserved:1 additional:1 somewhat:1 care:1 ey:1 diagonality:1 signal:2 multiple:2 full:3 infer:1 cross:2 long:1 dept:2 devised:1 molecular:2 equally:1 coded:1 a1:6 award:1 impact:1 prediction:17 variant:1 expectation:3 sometimes:1 represent:2 normalization:1 achieved:1 cell:9 microscope:1 c1:33 addition:1 hurdle:1 background:1 microarray:11 crucial:1 exhibited:1 tend:1 undergo:1 mature:2 hybridization:6 jordan:1 presence:2 exceed:1 concerned:1 decent:1 affect:2 fit:3 identified:1 regarding:1 lesser:1 br:1 microarrays:3 listening:1 expression:13 specialization:1 six:7 utility:1 motivated:1 generally:3 detailed:1 clear:1 amount:3 slide:5 band:1 morris:1 dna:11 northern:1 estimated:2 correctly:3 per:1 skeletal:1 express:2 four:3 liv:1 prevent:1 kept:1 year:1 sum:2 sti:1 oligonucleotide:1 parameterized:1 place:1 spleen:1 splicing:20 prefer:1 mined:1 durbin:1 photosensitive:1 scanned:1 occur:1 constraint:4 extremely:1 structured:1 combination:3 flake:1 across:3 pan:3 em:8 rev:1 making:3 s1:3 outlier:8 intuitively:1 invariant:1 heart:1 taken:1 equation:5 mutually:1 remains:1 previously:1 turn:1 mechanism:1 needed:1 studying:1 ofer:1 junction:10 available:1 probe:34 away:2 alternative:24 original:1 top:3 assumes:1 ensure:2 graphical:3 xc:3 ghahramani:1 approximating:1 added:2 exclusive:1 rt:14 diagonal:3 evolutionary:1 blot:1 exhibit:3 gradient:2 thank:1 oa:2 extent:1 considers:1 code:2 relationship:1 gel:2 ratio:2 regulation:5 setup:2 negative:2 design:2 allowing:1 observation:4 enabling:1 finite:1 descent:1 truncated:3 hinton:1 intensity:6 inferred:3 introduced:1 required:1 kl:1 optimized:1 baysian:1 learned:4 established:1 address:1 able:1 suggested:1 oc1:1 pcr:14 green:1 event:16 critical:1 difficulty:2 ranked:2 predicting:1 indicator:1 technology:2 nonsense:1 carried:1 washed:1 genomics:1 text:1 prior:2 understanding:1 discovery:1 contributing:1 relative:3 plant:1 expect:2 fluorescent:4 proportional:1 nucleus:1 affine:1 sufficient:1 consistent:1 elsewhere:2 summary:1 accounted:1 surprisingly:1 keeping:1 free:4 silence:1 guide:1 xc1:1 pulled:1 wide:2 saul:1 intestine:1 sparse:1 distributed:1 calculated:1 vari:1 genome:3 stranded:1 approximate:1 gene:17 global:1 overfitting:1 assumed:2 xi:3 alternatively:1 latent:2 regulatory:1 table:1 additionally:2 learn:2 nature:3 molecule:2 flank:1 electric:1 diag:1 sp:1 linearly:1 s2:4 noise:7 profile:5 allowed:1 facilitating:1 complementary:1 body:1 positively:1 site:2 referred:3 en:1 ational:1 candidate:2 third:1 abundance:13 splice:6 specific:5 showing:1 intron:4 er:1 r2:3 multitude:1 a3:3 intractable:1 false:5 adding:1 importance:1 te:1 labelling:1 justifies:1 easier:1 timothy:1 rocke:1 forming:1 strand:3 adjustment:1 nserc:1 scalar:1 binding:1 oc2:1 corresponds:1 extracted:3 conditional:1 marked:2 viewed:1 quantifying:1 labelled:2 included:4 determined:1 except:1 typical:1 called:1 specie:1 experimental:1 e:1 qun:1 meaningful:2 exception:1 indicating:1 select:1 evaluate:2 phenomenon:1 |
1,848 | 2,679 | Message Errors in Belief Propagation
Alexander T. Ihler, John W. Fisher III, and Alan S. Willsky
Department of Electrical Engineering and Computer Science
Massachusetts Institute of Technology
[email protected], [email protected], [email protected]
Abstract
Belief propagation (BP) is an increasingly popular method of performing approximate inference on arbitrary graphical models. At times,
even further approximations are required, whether from quantization or
other simplified message representations or from stochastic approximation methods. Introducing such errors into the BP message computations
has the potential to adversely affect the solution obtained. We analyze
this effect with respect to a particular measure of message error, and show
bounds on the accumulation of errors in the system. This leads both to
convergence conditions and error bounds in traditional and approximate
BP message passing.
1
Introduction
Graphical models and message-passing algorithms defined on graphs are a growing field of
research. In particular, the belief propagation (BP, or sum-product) algorithm has become
a popular means of solving inference problems exactly or approximately. One part of
its appeal is its optimality for tree-structured graphical models (models which contain no
loops). However, its is also widely applied to graphical models with cycles. In these cases
it may not converge, and if it does its solution is approximate; however in practice these
approximations are often good. Recently, further justifications for loopy belief propagation
have been developed, including a few convergence results for graphs with cycles [1?3].
The approximate nature of loopy BP is often a more than acceptable price for efficient inference; in fact, it is sometimes desirable to make additional approximations. There may be
a number of reasons for this?for example, when exact message representation is computationally intractable, the messages may be approximated stochastically [4] or deterministically by discarding low-likelihood states [5]. For BP involving continuous, non-Gaussian
potentials, some form of approximation is required to obtain a finite parametrization for
the messages [6?8]. Additionally, graph simplification by edge removal may be regarded
as a coarse form of message approximation. Finally, one may wish to approximate the
messages and reduce their representation size for another reason?to decrease the communications required for distributed inference applications. In a distributed environment, one
may approximate the transmitted message to reduce its representational cost [9], or discard
it entirely if it is deemed ?sufficiently similar? to the previously sent version [10]. This may
significantly reduce the amount of communication required.
Given that message approximation may be desirable, we would like to know what effect
the introduced errors have on our overall solution. To characterize the effect in graphs
with cycles, we analyze the deviation from a solution given by ?exact? loopy BP (not, as is
typically considered, the deviation of loopy BP from the true marginal distributions). In the
process, we also develop some results on the convergence of loopy BP. Section 3 describes
the major themes of the paper; but first we provide a brief summary of belief propagation.
2
Graphical Models and Belief Propagation
Graphical models provide a convenient means of representing conditional independence
relations among large numbers of random variables. Specifically, each node s in a graph
is associated with a random variable xs , while the set of edges E is used to describe the
conditional dependency structure of the variables. A distribution satisfies the conditional
independence relations specified by an undirected graph if it factors into a product of potential functions ? defined on the cliques (fully-connected subsets) of the graph; the converse
is also true if p(x) is strictly positive [11]. Here we consider graphs with at most pairwise
interactions (a typical assumption in BP), where the distribution factors according to
Y
Y
p(x) =
?st (xs , xt )
?s (xs )
(1)
s
(s,t)?E
The goal of belief propagation [12], or BP, is to compute the marginal distribution p(xt ) at
each node t. BP takes the form of a message-passing algorithm between nodes, expressed
in terms of an update to the outgoing message from each node t to each neighbor s in terms
of the (previous iteration?s) incoming messages from t?s neighbors ?t ,
Z
Y
mts (xs ) ? ?ts (xt , xs )?t (xt )
mut (xt )dxt
(2)
u??t \s
Typically each message is normalized so as to integrate to unity (and we assume that such
normalization is possible). At any iteration, one may calculate the belief at node t by
Y
Mti (xt ) ? ?t (xt )
miut (xt )
(3)
u??t
For tree-structured graphical models, belief propagation can be used to efficiently perform
exact marginalization. Specifically, the iteration (2) converges in a finite number of iterations (at most the length of the longest path in the graph), after which the belief (3) equals
the correct marginal p(xt ). However, as observed by [12], one may also apply belief propagation to arbitrary graphical models by following the same local message passing rules
at each node and ignoring the presence of cycles in the graph; this procedure is typically
referred to as ?loopy? BP.
For loopy BP, the sequence of messages defined
by (2) is not guaranteed to converge to a fixed
point after any number of iterations. Under relatively mild conditions, one may guarantee the existence of fixed points [13]. However, they may
not be unique, nor are the results exact (the belief Mti does not converge to the true marginal).
In practice however the procedure often arrives at
a reasonable set of approximations to the correct
marginal distributions.
1
1
3
2
4
2
3
4
4
1
3
2
3
1
1
4
1
2
Figure 1: For a graph with cycles, one
may show an equivalence between n iterations of loopy BP and the n-level
computation tree (shown here for n = 3
and rooted at node 1; example from [2]).
It is sometimes convenient to think of loopy BP
in terms of its computation tree [2]. The n-level
computation tree rooted at some node t is a treestructured ?unrolling? of the graph, so that n iterations of loopy BP on the original graph
is equivalent at the node t to exact inference on the computation tree. An example of this
structure is shown in Figure 1.
log m/m
?
m(x)
m(x)
?
(a)
} log d (e)
(b)
Figure 2: (a) A message m(x), solid, and its approximation m(x),
?
dashed. (b) Their log-ratio
log m(x)/m(x);
?
log d (e) characterizes their similarity by measuring the error?s dynamic range.
3
Overview of Results
To orient the reader, we lay out the order and general results which are obtained in this paper. We begin by considering multiplicative error functions which describe the difference
between a ?true? message m(x) (typically meaning consistent with some BP fixed-point)
and some approximation m(x)
?
= m(x) ? e(x). We apply a particular functional measure
d (e) (defined below) and show how this measure behaves with respect to the BP equations (2) and (3). When applied to traditional BP, this results in a novel sufficient condition
for its convergence to a unique solution, specifically
X d (?ut )2 ? 1
max
< 1,
(4)
2
(s,t)?E
d (?ut ) + 1
u??t \s
and may be further improved in most cases. The condition (4) is shown to be slightly
stronger than the sufficient condition given in [2]. More importantly, however, the method
in which it is derived allows us to generalize to many other situations:
? The condition (4) is easily improved for graphs with irregular geometry or potential
strengths
? The method also provides a bound on the distance between any two BP fixed points.
? The same methodology may be applied to the case of quantized or otherwise approximated messages, yielding bounds on the ensuing error (our original motivation).
? By regarding message errors as a stochastic process and applying a few additional
assumptions, a similar analysis obtains alternate, tighter estimates (though not
necessarily bounds) of performance.
4
Message Approximations
In order to discuss the effects and propagation of errors introduced to the BP messages,
we first require a measure of the difference between two messages. Although there are
certainly other possibilities, it is very natural to consider the message deviations (which we
denote ets ) to be multiplicative, or additive in the log-domain, and examine a measure of
the error?s dynamic range:
p
m
? ts (xs ) = mts (xs )ets (xs )
d (ets ) = max ets (a)/ets (b)
(5)
a,b
Then, we have that mts (x) = m
? ts (x)?x if and only if log d (ets ) = 0. This measure may
also be related to more traditional error measures, including an absolute error on log m(x),
a floating-point precision on m(x), and the Kullback-Leibler divergence D(m(x)km(x));
?
for details, see [14]. In this light our analysis of message approximation (Section 5.3)
may be equivalently regarded as a statement about the required precision for an accurate
implementation of loopy BP. Figure 2 shows an example message m(x) and approximation
m(x)
?
along with their associated error e(x).
To facilitate our analysis, we split the message update operation (2) into two parts. In the
first, we focus on the message products
Y
Y
Mts (xt ) ? ?t (xt )
mut (xt )
Mt (xt ) ? ?t (xt )
mut (xt )
(6)
u??t \s
u??t
where as usual, the proportionality constant is chosen to normalize M . We show the message error metric is (sub-)additive, i.e. that the errors in each incoming message (at most)
add in their impact on M . The secondZoperation is the message convolution
mts (xs ) ?
?ts (xt , xs )Mts (xt )dxt
(7)
where M is a normalized message or product of messages. We demonstrate a level of
contraction, that is, the approximation of mts is measurably better than the approximation
of Mts used to construct it.
We use the convention that lowercase quantities (mts , ets , . . .) refer to messages and message errors, while uppercase ones (Mts , Ets , Mt , . . .) refer to products of messages or
errors?all incoming messages to node t (Mt and Et ), or all except the one from s (Mts
and Ets ). Due to space constraints, many omitted details and proofs can be found in [14].
4.1
Additivity and Error Contraction
The log of (5) is sub-additive, since
for several
incoming
? ut (x)} we have
?
?
?Y messages
? X{m
?
log d (Ets ) = log d Mts /Mts = log d
eut ?
log d (eut )
(8)
We may also derive a minimum rate of contraction on the errors. We consider the message
from t to s; since all quantities in this section relate to mts and Mts we suppress the
subscripts. The error measure Rd (e) is given by
R
?(xt , a)M (xt )E(xt )dxt
?(xt , b)M (xt )dxt
2
2
R
R
d (e) = d (m/m)
?
= max
?
(9)
a,b
?(xt , a)M (xt )dxt
?(xt , b)M (xt )E(xt )dxt
subject to certain constraints, such
R as positivity
R of the messages and potentials. Since
?f, g > 0,
f (x) dx / g(x) dx ? max f (x)/g(x)
(10)
x
log d (e) ?
we can directly obtain the two bounds:
2
2
2
4
(11)
d (e) ? d (E) and d (e) ? d (?)
where we have extended the measure d (?) to
functions of two variables (describing a minimum rate of mixing across the potential) by
log d (E)
2
log d (?)
?(a, b)
2
.
(12)
d (?) = max
d(E)+1
a,b,c,d ?(c, d)
log d(?)
d(?) +d(E)
However, with some work one may show [14] the
stronger measure of contraction,
log d (E) ?
2
d (?) d (E) + 1
.
(13)
d (e) ?
2
Figure 3: Bounds on the error output
d (?) + d (E)
2
2
d (e) as a function of the error in the
Sketch of proof: While the full proof is rather involved,
product of incoming messages d (E).
we outline the procedure here. First, use (10) to show
that the maximum of (9) given d (?) is attained by potentials of the form ?(x, a) ? 1 + K?A (x)
and ?(x, b) ? 1 + K?B (x), where K = d (?)2 ? 1 and ?A and ?B take
R on only values {0, 1},
along
with
a
similar
form
for
E(x).
Then
define
the
variables
M
=
M (x)?A (x), MAE =
A
R
M (x)?A (x)?E (x), etc., and optimize given the constraints 0 ? MA , MB , ME ? 1, MAE ?
min[MA , ME ], and MBE ? max[0, ME ? (1 ? MB )] (where the last constraint arises from the
fact that ME + MB ? MBE ? 1). Simplifying and taking the square root yields (13).
The bound (13) is shown in Figure 3; note that it improves both error bounds (11), shown
as straight lines. In the next section, we use (8)-(13) to analyze the behavior of loopy BP.
5
Implications in Graphs with Cycles
We begin by examining loopy BP with exact message passing, using the previous results
to derive a new sufficient condition for convergence to a unique fixed point. When this
condition is not satisfied, we instead obtain a bound on the relative distances between any
two fixed points of the loopy BP equations. We then consider the effect of introducing
additional errors into the messages passed at each iteration, showing sufficient conditions
for this operation to converge, and a bound on the resulting error from exact loopy BP.
5.1
Convergence of Loopy BP & Fixed Point Distance
Tatikonda and Jordan [2] showed that the convergence and fixed points of loopy BP may
be considered in terms of a Gibbs measure on the graph?s computation tree, implying that
loopy BP is guaranteed to converge if the graph satisfies Dobrushin?s condition [15]. Dobrushin?s condition is a global measure and difficult to verify; given in [2] is a sufficient
condition (often called Simon?s condition):
X
max
log d (?ut ) < 1
(14)
t
u??t
where d (?) is defined as in (12). Using the previous section?s analysis, we may argue
something slightly stronger. Let us take the ?true? messages mts to be any fixed point of
BP, and ?approximate? them at each iteration by performing loopy BP from some arbitrary
initial conditions. Now suppose that the largest message-product error log d (Eut ) in any
node u with parent t at level i of the computation tree (corresponding to iteration n ? i out
of n total iterations of loopy BP) is bounded above by some constant log ?i . Note that this
2
is trivially true (at any i) for the constant log ?i = max(u,t)?E |?t | log d (?ut ) . Now, we
may bound d (Ets ) at any replicate of node t with parent s on level i ? 1 of the tree by
2
X
d (?ut ) ?i + 1
log d (Ets ) ? gts (log ?i ) =
log
.
(15)
2
d (?ut ) + ?i
u??t \s
and we may define log ?i?1 = maxt,s gts (log ?i ) to bound the error at level i?1. Loopy BP
will converge if the sequence ?i , ?i?1 , . . . is strictly decreasing for all ? > 1, i.e. gts (z) < z
0
00
for all z > 0. This is guaranteed by the conditions gts (0) = 0, gts
(0) < 1 and gts
(z) < 0.
0
The first is easy to show, the third can be verified by algebra, and the condition gts
(0) < 1
can be rewritten to give the convergence criterion
X d (?ut )2 ? 1
<1
(16)
max
2
(s,t)?E
d (?ut ) + 1
u??t \s
We may relate (16) to Simon?s condition (14) by expanding the set ?t \ s to the larger ?t
2
and noting that log x ? xx2 ?1
+1 for all x ? 1 with equality as x ? 1. Doing so, we see
that Simon?s condition is sufficient to guarantee (16), but that (16) may be true (implying
convergence) when Simon?s condition is not satisfied. The improvement over Simon?s condition becomes negligible as connectivity increases (assuming the graph has approximately
equal-strength potentials), but can be significant for low connectivity. For example, if the
graph consists of a single loop then each node t has at most two neighbors. In this case,
the contraction (16) tells us that the outgoing message in either direction is always closer
to the BP fixed point than the incoming message. Thus we obtain the result of [1], that (for
finite-strength potentials) BP always converges to a unique fixed point on graphs containing
a single loop. Simon?s condition, on the other hand, is too loose to demonstrate this fact.
If the condition (16) is not satisfied, then the sequence {?i } is not always decreasing and
there may be multiple fixed points. In this case, the sequence {?i } as defined will decrease
until it reaches the largest value ? such that maxts gts (log ?) = log ?. Since the choice of
initialization was arbitrary, we may opt to initialize to any other fixed point, and observe
that the difference Et between these two fixed point beliefs is bounded by
2
X
d (?ut ) ? + 1
log d (Et ) ?
log
(17)
2
d (?ut ) + ?
u??t
log d (Et )
10
8
6
Simple bound, grids (a) and (b)
Nonuniform bound, grid (a)
Nonuniform bound, grid (b)
Simons condition
4
2
0
0
(a)
(b)
??
0.5
1
1.5
2
2.5
(c)
Figure 4: Two small (5 ? 5) grids, with (a) all equal-strength potentials log d (?)2 = ? and (b)
several weaker ones (log d (?)2 = .5?, thin lines). The methods described provide bounds (c) on the
distance between any two fixed points as a function of potential strength ?, all of which improve on
Simon?s condition. See text for details.
Thus, the fixed points of BP lie in some potentially small set. If log ? is small (the condition (16) is nearly satisfied) then although we cannot guarantee convergence to a unique
fixed point, we can guarantee that every fixed point and our estimate are all mutually close
(in a log-ratio sense).
5.2
Improving the Bounds by Path-counting
If we are willing to put a bit more effort into our bound-computation, we may be able to
improve it.In particular the proofs of (16)-(17) assume that, as a message error propagates
through the graph, repeated convolution with only the strongest set of potentials is possible.
But often even if the worst potentials are quite strong, every cycle which contains them
also contains several weaker potentials. Using an iterative algorithm much like BP itself,
we may obtain a more globally aware estimate of error propagation.
Let us consider a message-passing procedure (potentially
offline) in which node
? performed
?
i
on the message error d eits at iteration i to its neighbor s.
t passes a (scalar) bound ?ts
2
1
The bound may be initialized to ?ts
= d (?ts ) , and the next iteration?s (updated) outgoing
bound is given by the pair of equations
2
X
d (?ts ) ?its + 1
i+1
i
i
log ?ts
= log
log
?
=
log ?ut
(18)
ts
2
d (?ts ) + ?its
u??t \s
Here, as in Section 5.1, ?its bounds the error d (Ets ) in the product of incoming messages.
i
If (18) converges to log ?ts
? 0 for all t, s P
we may guarantee a unique fixed point for
i
loopy BP; if not, we may compute log ?it = ?t log ?ut
to obtain a bound on the belief
error at each node t. If every node is identical (same number of neighbors, same potential
strengths) this yields the same bound as (17); however, if the graph or potential strengths are
inhomogeneous it provides a strictly stronger bound on loopy BP convergence and errors.
This situation is illustrated in Figure 4?we specify two 5?5 grids in terms of their potential
strengths and compute bounds on the log-range of their fixed point beliefs. (While potential
strength does not completely specify the graphical model, it is sufficient for all the bounds
2
considered here.) One grid (a) has equal-strength potentials log d (?) = ?, while the other
has many weaker potentials (?/2). The worst-case bounds are the same (since both have
a node with four strong neighbors), shown as the solid curve in (c). However, the dashed
curves show the estimate of (18), which improves only slightly for the strongly coupled
graph (a) but considerably for the weaker graph (b). All three bounds give considerably
more information than Simon?s condition (dotted vertical line).
5.3
Introducing additional errors
As discussed in the introduction, we may wish to introduce or allow additional errors in our
messages at each stage, in order to improve the computational or communication efficiency
of the algorithm. This may be the result of an actual distortion imposed on the message
(perhaps to decrease its complexity, for example quantization), from censoring the message
update (reusing the message from the previous iteration) when the two are sufficiently
similar, or from approximating or quantizing the model parameters (potential functions).
Any of these additional errors can be easily incorporated into our framework.
If at each iteration, we introduce an additional (perhaps stochastic) error to each message
which has a dynamic range bounded by some constant ?, the relationships of (18) become
2
X
d (?ts ) ?its + 1
i+1
i
i
log ?ts
= log
+
log
?
log
?
=
log ?ut
(19)
ts
2
d (?ts ) + ?its
u??t \s
and gives a bound on the steady-state error (distance from a fixed point) in the system.
5.4
Stochastic Analysis
Unfortunately, the above bounds are often pessimistic compared to actual performance. By
treating the perturbations as stochastic we may obtain a more realistic estimate (though no
longer a strict bound) on the resulting error. Specifically, let us describe the error func2
tions log ets (xs ) for each xs as a random variable with mean zero and variance ?ts
. By
assuming that the errors inP
each incoming message are uncorrelated, we obtain additivity
2
of their variances: ?2ts = u??t \s ?ut
. The assumption of uncorrelated errors is clearly
questionable since propagation around loops may couple the incoming message errors, but
is common in quantization analysis, and we shall see that it appears reasonable in practice.
We would also like to estimate the contraction of variance incurred in the convolution step.
We may do so by applying a simple sigma-point quadrature (?unscented?) approximation [16], in which the standard deviation of the convolved function mts (xs ) is estimated
by applying the same nonlinearity (13) to the standard deviation of the error on the incoming product Mts . Thus, similarly to (18) and (19), we have
!2
?
2
X
d (?ts ) ?ts + 1
2
2
+ (log ?)2
(log ?ts )2 =
?ut
(20)
?ts = log
2
?ts + d (?ts )
u??t \s
The steady-state
solution of (20) yields an estimate of the variances of the log-belief log pt
P
2
by ?t2 = u??t ?ut
; this estimate is typically much smaller than the bound (18) due to the
strict sub-additive relationship between the standard deviations. Although it is not a bound,
using a Chebyshev-like argument we may conclude that, for example, the 2?t distance will
be greater than the typical errors observed in practice.
6
Experiments
We demonstrate the error bounds for perturbed messages with a set of Monte Carlo trials.
In particular, for each trial we construct a binary-valued 5 ? 5 grid with uniform potential strengths, which are either (1) all positively correlated, or (2) randomly chosen to be
positively or negatively correlated (equally likely); we also assign random single-node potentials to each xs . We then run a quantized version of BP, rounding each log-message
to discrete values separated by 2 log ? (ensuring that the newly introduced error satisifies
d (e) ? ?). Figure 5 shows the maximum belief error in each of 100 trials of this procedure
for various values of ?.
Also shown are the bound on belief error developed in Section 5.3 and the 2? estimate
computed assuming uncorrelated message errors. As can be seen, the stochastic estimate
is often a much tighter, more accurate assessment of error, but it does not possess the
same strong theoretical guarantees. Since, as observed in analysis of quantization and
stability in digital filtering [17], the errors introduced by quantization are typically close to
independent, the assumptions of the stochastic estimate are reasonable and empirically we
observe that the estimate and actual errors behave similarly.
max log d (Et )
max log d (Et )
Strict bound
Stochastic estimate
Positive corr. potentials
Mixed corr. potentials
1
10
0
10
10
10
10
1
2
log ? ?
3
3
10
10
2
10
1
2
0
10
(a) log d (?) = .25
10 1
100
10 -1
10 -2
log ? ?
10 -3 -3
10
10-2
10-1
100
2
(b) log d (?) = 1
Figure 5: Maximum belief errors incurred as a function of the quantization error. The scatterplot
indicates the maximum error measured in the graph for each of 200 Monte Carlo runs; this is strictly
bounded above by the solution of (18), solid, and bounded with high probability (assuming uncorrelated errors) by (20), dashed.
7
Conclusions
We have described a particular measure of distortion on BP messages and shown that it is
sub-additive and measurably contractive, leading to sufficient conditions for loopy BP to
converge to a unique fixed point. Furthermore, this enables analysis of quantized, stochastic, or other approximate forms of BP, yielding sufficient conditions for convergence and
bounds on the deviation from exact message passing. Assuming the perturbations are uncorrelated can often give tighter estimates of the resulting error. For additional details as
well as some further consequences and extensions, see [14].
The authors would like to thank Erik Sudderth, Martin Wainwright, Tom Heskes, and Lei Chen for
many helpful discussions. This research was supported in part by MIT Lincoln Laboratory under
Lincoln Program 2209-3023 and by ODDR&E MURI through ARO grant DAAD19-00-0466.
References
[1] Y. Weiss. Correctness of local probability propagation in graphical models with loops. Neural
Computation, 12(1), 2000.
[2] S. Tatikonda and M. Jordan. Loopy belief propagation and gibbs measures. In UAI, 2002.
[3] T. Heskes. On the uniqueness of loopy belief propagation fixed points. To appear in Neural
Computation, 2004.
[4] D. Koller, U. Lerner, and D. Angelov. A general algorithm for approximate inference and its
application to hybrid Bayes nets. In UAI 15, pages 324?333, 1999.
[5] J. M. Coughlan and S. J. Ferreira. Finding deformable shapes using loopy belief propagation.
In ECCV 7, May 2002.
[6] E. B. Sudderth, A. T. Ihler, W. T. Freeman, and A. S. Willsky. Nonparametric belief propagation.
In CVPR, 2003.
[7] M. Isard. PAMPAS: Real?valued graphical models for computer vision. In CVPR, 2003.
[8] T. Minka. Expecatation propagation for approximate bayesian inference. In UAI, 2001.
[9] A. T. Ihler, J. W. Fisher III, and A. S. Willsky. Communication-constrained inference. Technical
Report TR-2601, Laboratory for Information and Decision Systems, 2004.
[10] L. Chen, M. Wainwright, M. Cetin, and A. Willsky. Data association based on optimization in
graphical models with application to sensor networks. Submitted to Mathematical and Computer Modeling, 2004.
[11] P. Clifford. Markov random fields in statistics. In G. R. Grimmett and D. J. A. Welsh, editors,
Disorder in Physical Systems, pages 19?32. Oxford University Press, Oxford, 1990.
[12] J. Pearl. Probabilistic Reasoning in Intelligent Systems. Morgan Kaufman, San Mateo, 1988.
[13] J. S. Yedidia, W. T. Freeman, and Y. Weiss. Constructing free energy approximations and
generalized belief propagation algorithms. Technical Report 2004-040, MERL, May 2004.
[14] A. T. Ihler, J. W. Fisher III, and A. S. Willsky. Message errors in belief propagation. Technical
Report TR-2602, Laboratory for Information and Decision Systems, 2004.
[15] Hans-Otto Georgii. Gibbs measures and phase transitions. Studies in Mathematics. de Gruyter,
Berlin / New York, 1988.
[16] S. Julier and J. Uhlmann. A general method for approximating nonlinear transformations of
probability distributions. Technical report, RRG, Dept. of Eng. Science, Univ. of Oxford, 1996.
[17] A. Willsky. Relationships between digital signal processing and control and estimation theory.
Proc. IEEE, 66(9):996?1017, September 1978.
| 2679 |@word mild:1 trial:3 version:2 stronger:4 replicate:1 proportionality:1 km:1 willing:1 contraction:6 simplifying:1 eng:1 tr:2 solid:3 initial:1 contains:2 dx:2 john:1 realistic:1 additive:5 shape:1 enables:1 treating:1 update:3 implying:2 isard:1 parametrization:1 coughlan:1 coarse:1 provides:2 node:18 quantized:3 mathematical:1 along:2 become:2 consists:1 introduce:2 pairwise:1 behavior:1 nor:1 growing:1 examine:1 freeman:2 globally:1 decreasing:2 actual:3 considering:1 unrolling:1 becomes:1 begin:2 bounded:5 what:1 kaufman:1 developed:2 finding:1 transformation:1 guarantee:6 every:3 questionable:1 exactly:1 ferreira:1 control:1 converse:1 grant:1 appear:1 positive:2 negligible:1 engineering:1 local:2 cetin:1 consequence:1 ets:14 oxford:3 subscript:1 path:2 approximately:2 initialization:1 mateo:1 equivalence:1 range:4 contractive:1 unique:7 eut:3 practice:4 daad19:1 procedure:5 significantly:1 convenient:2 inp:1 cannot:1 close:2 put:1 applying:3 accumulation:1 equivalent:1 optimize:1 imposed:1 disorder:1 rule:1 regarded:2 importantly:1 stability:1 justification:1 updated:1 pt:1 suppose:1 exact:8 approximated:2 lay:1 muri:1 observed:3 electrical:1 worst:2 calculate:1 cycle:7 connected:1 decrease:3 environment:1 complexity:1 dynamic:3 solving:1 algebra:1 negatively:1 efficiency:1 completely:1 easily:2 various:1 additivity:2 separated:1 univ:1 describe:3 monte:2 tell:1 quite:1 widely:1 larger:1 valued:2 distortion:2 cvpr:2 otherwise:1 otto:1 statistic:1 think:1 itself:1 sequence:4 quantizing:1 net:1 aro:1 interaction:1 product:9 mb:3 loop:5 mixing:1 representational:1 lincoln:2 deformable:1 normalize:1 pampas:1 convergence:12 parent:2 converges:3 tions:1 derive:2 develop:1 measured:1 strong:3 convention:1 direction:1 inhomogeneous:1 correct:2 stochastic:9 func2:1 require:1 assign:1 opt:1 tighter:3 pessimistic:1 strictly:4 extension:1 unscented:1 sufficiently:2 considered:3 around:1 major:1 omitted:1 uniqueness:1 estimation:1 proc:1 tatikonda:2 uhlmann:1 treestructured:1 largest:2 correctness:1 mit:4 clearly:1 sensor:1 gaussian:1 always:3 rather:1 derived:1 focus:1 longest:1 improvement:1 likelihood:1 indicates:1 sense:1 helpful:1 inference:8 lowercase:1 typically:6 relation:2 koller:1 overall:1 among:1 constrained:1 initialize:1 marginal:5 field:2 equal:4 construct:2 aware:1 identical:1 nearly:1 thin:1 t2:1 report:4 intelligent:1 few:2 randomly:1 lerner:1 divergence:1 mut:3 floating:1 geometry:1 welsh:1 phase:1 message:67 possibility:1 certainly:1 arrives:1 yielding:2 light:1 uppercase:1 implication:1 accurate:2 edge:2 closer:1 tree:9 initialized:1 theoretical:1 merl:1 modeling:1 measuring:1 loopy:27 cost:1 introducing:3 deviation:7 subset:1 uniform:1 examining:1 rounding:1 too:1 characterize:1 dependency:1 perturbed:1 considerably:2 st:1 csail:1 probabilistic:1 connectivity:2 clifford:1 satisfied:4 containing:1 positivity:1 adversely:1 stochastically:1 leading:1 reusing:1 potential:25 de:1 multiplicative:2 root:1 performed:1 analyze:3 characterizes:1 doing:1 bayes:1 simon:9 square:1 variance:4 efficiently:1 yield:3 generalize:1 bayesian:1 carlo:2 straight:1 submitted:1 strongest:1 reach:1 energy:1 involved:1 minka:1 associated:2 ihler:5 proof:4 couple:1 newly:1 massachusetts:1 popular:2 ut:17 improves:2 appears:1 attained:1 methodology:1 specify:2 improved:2 tom:1 wei:2 though:2 strongly:1 furthermore:1 stage:1 until:1 sketch:1 hand:1 nonlinear:1 assessment:1 propagation:20 perhaps:2 lei:1 facilitate:1 effect:5 contain:1 true:7 verify:1 normalized:2 equality:1 leibler:1 laboratory:3 illustrated:1 mbe:2 rooted:2 steady:2 criterion:1 generalized:1 outline:1 demonstrate:3 reasoning:1 meaning:1 novel:1 recently:1 common:1 behaves:1 functional:1 mt:21 empirically:1 overview:1 physical:1 discussed:1 association:1 julier:1 mae:2 refer:2 significant:1 gibbs:3 rd:1 trivially:1 grid:7 similarly:2 heskes:2 mathematics:1 nonlinearity:1 han:1 similarity:1 longer:1 etc:1 add:1 something:1 showed:1 discard:1 certain:1 binary:1 transmitted:1 minimum:2 additional:8 greater:1 seen:1 morgan:1 converge:7 signal:1 dashed:3 full:1 desirable:2 multiple:1 alan:1 technical:4 equally:1 impact:1 ensuring:1 involving:1 vision:1 metric:1 iteration:15 sometimes:2 normalization:1 irregular:1 sudderth:2 georgii:1 posse:1 pass:1 strict:3 subject:1 sent:1 undirected:1 gts:8 jordan:2 presence:1 noting:1 counting:1 iii:3 split:1 easy:1 affect:1 independence:2 marginalization:1 reduce:3 regarding:1 chebyshev:1 whether:1 passed:1 effort:1 passing:7 york:1 amount:1 nonparametric:1 dotted:1 estimated:1 discrete:1 shall:1 four:1 verified:1 graph:25 mti:2 satisifies:1 sum:1 orient:1 run:2 reasonable:3 reader:1 decision:2 acceptable:1 bit:1 entirely:1 bound:39 guaranteed:3 simplification:1 strength:11 constraint:4 bp:44 argument:1 optimality:1 min:1 performing:2 relatively:1 martin:1 department:1 structured:2 according:1 alternate:1 describes:1 slightly:3 increasingly:1 across:1 unity:1 smaller:1 computationally:1 equation:3 mutually:1 previously:1 discus:1 describing:1 loose:1 know:1 operation:2 rewritten:1 yedidia:1 apply:2 observe:2 grimmett:1 convolved:1 existence:1 original:2 graphical:12 approximating:2 quantity:2 usual:1 traditional:3 september:1 distance:6 thank:1 berlin:1 ensuing:1 me:4 argue:1 reason:2 willsky:7 assuming:5 erik:1 length:1 relationship:3 ratio:2 equivalently:1 difficult:1 unfortunately:1 statement:1 relate:2 potentially:2 sigma:1 suppress:1 implementation:1 perform:1 vertical:1 convolution:3 markov:1 finite:3 t:24 behave:1 situation:2 extended:1 communication:4 incorporated:1 nonuniform:2 perturbation:2 arbitrary:4 introduced:4 pair:1 required:5 specified:1 pearl:1 able:1 below:1 program:1 including:2 max:11 belief:25 wainwright:2 natural:1 hybrid:1 xx2:1 representing:1 improve:3 technology:1 brief:1 deemed:1 coupled:1 text:1 removal:1 relative:1 fully:1 dxt:6 mixed:1 filtering:1 digital:2 integrate:1 incurred:2 sufficient:9 consistent:1 propagates:1 editor:1 uncorrelated:5 maxt:1 eccv:1 censoring:1 summary:1 supported:1 last:1 free:1 offline:1 weaker:4 allow:1 institute:1 neighbor:6 taking:1 absolute:1 distributed:2 curve:2 transition:1 author:1 san:1 simplified:1 approximate:10 obtains:1 kullback:1 clique:1 global:1 incoming:10 uai:3 conclude:1 continuous:1 iterative:1 additionally:1 nature:1 expanding:1 ignoring:1 improving:1 angelov:1 necessarily:1 constructing:1 domain:1 motivation:1 repeated:1 quadrature:1 positively:2 measurably:2 referred:1 precision:2 sub:4 theme:1 deterministically:1 wish:2 lie:1 third:1 discarding:1 xt:27 showing:1 appeal:1 x:14 intractable:1 quantization:6 scatterplot:1 corr:2 chen:2 likely:1 expressed:1 scalar:1 satisfies:2 ma:2 gruyter:1 conditional:3 goal:1 price:1 fisher:4 specifically:4 typical:2 except:1 called:1 total:1 arises:1 alexander:1 dobrushin:2 dept:1 outgoing:3 correlated:2 |
1,849 | 268 | 372
Touretzky and Wheeler
A Computational Basis for Phonology
David S. Touretzky
School of Computer Science
Carnegie Mellon University
Pittsburgh, PA 15213
Deirdre W. Wheeler
Department of Linguistics
University of Pittsburgh
Pittsburgh, PA 15260
ABSTRACT
The phonological structure of human languages is intricate, yet highly
constrained. Through a combination of connectionist modeling and
linguistic analysis, we are attempting to develop a computational basis
for the nature of phonology. We present a connectionist architecture
that performs multiple simultaneous insertion, deletion, and mutation
operations on sequences of phonemes, and introduce a novel additional
primitive, clustering. Clustering provides an interesting alternative to
both iterative and relaxation accounts of assimilation processes such as
vowel harmony. Our resulting model is efficient because it processes
utterances entirely in parallel using only feed-forward circuitry.
1 INTRODUCTION
Phonological phenomena can be quite complex, but human phonological behavior is
also highly constrained. Many operations that are easily learned by a perceptron-like
sequence mapping network are excluded from real languages. For example, as Pinker
and Prince (1988) point out in their critique of the Rumelhart and McClelland (1986)
verb learning model, human languages never reverse the sequence of segments in a
word, but this is an easy mapping for a network to learn. On the other hand, we note that
some phonological processes that are relatively common in human languages, such as
vowel harmony, appear difficult for a sequence-mapping architecture to learn. Why are
only certain types of sequence operations found in human languages, and not others? We
suggest that this is a reflection of the limitations of an underlying, genetically-determined,
specialized computing architecture. We are searching for this architecture.
A Computational Basis for Phonology
Our work was initially inspired by George Lakoff's theory of cognitive phonology
(Lakoff, 1988, 1989), which is in tum a development of the ideas of John Goldsmith
(to appear). Lakoff proposes a three-level representation scheme. The M (morphophonemic) level represents the underlying form of an utterance, the P (phonemic) level
is an intermediate form, and the F (phonetic) level is the derived surface form.
Lakoff uses a combination of inter-level mapping rules and intra-level well-formedness
conditions to specify the relationships between P- and F-Ievel representations and the
M-Ievel input. In a connectionist implementation, the computations performed by the
mapping rules are straightforward, but we find the well-formedness conditions troubling.
Goldsmith's proposal was that phonology is a goal-directed constraint satisfaction system that operates via parallel relaxation. He cites Smolensky's hannony theoryl Lakoff
has adopted this appeal to hannony theory in his description of how well-formedness
conditions could work.
In our model, we further develop the Goldmsith and Lakoff mapping scheme, but we reject
harmony-based well-formedness conditions for several reasons. First, harmony theory
involves simulated annealing search. The timing constraints of real nervous systems rule
out simulated annealing. Second, it is not clear how to construct an energy function for
a connectionist network that performs complex discrete phonological operations. Finally
there is our desire to explain why certain types of processes occur in human languages
and others do not. Harmony theory alone is too unconstrained for this purpose.
We have implemented a model called M3 p (for "Many Maps" Model of Phonology) that
allows us to account for virtually all of the phenomena in (Lakoff, 1989) using a tighUyconstrained, purely-feedforward computing scheme. In the next section we describe the
mapping matrix architecture that is the heart of M3 p . Next we give an example of an
iterative process, Yawelmani vowel harmony,2, which Lakoff models with a P-Ievel wellformedness condition. Such a condition would have to be implemented by relaxation
search for a "minimum energy state" in the P-Ievel representation, which we wish to
avoid. Finally we present our alternative approach to vowel harmony, using a novel
clustering mechanism that eliminates the need for relaxation.
2 THE MAPPING MATRIX ARCHITECTURE
Figure 1 is an overview of our "many maps" model. M-P constructions compute how
to go from the M-Ievel representation of an utterance to the P-Ievel representation. The
derivation is described as a set of explicit changes to the M-Ievel string. M-P constructions read the segments in the M-Ievel buffer and write the changes, phrased as mutation,
deletion, and insertion requests, into slots of a buffer called P-deriv. The M-Ievel and
P-deriv buffers are then read by the M-P mapping matrix, which produces the P-Ievel
representation as its output. The process is repeated at the next level, with P-F constructions writing changes into an F-deriv buffer, and a P-F map deriving an F-Ievel
1Srnolensky' s "hannony theory" should not be confused with the linguistic phenomenon of "vowel hannony."
2Yawelmani is a dialect of Yokuts, an American Indian language from California. Our Yawelmani data is
drawn from Kenstowicz and Kisseberth (1979), as is Lakoff's.
373
374
Touretzky and Wheeler
M-Level Buffer
I
M-P Constructions
P-Level Buffer
I
P-F Constructions
-----4~~ F-Level Buffer
I
Canonicalization
I
Surface Phonetic ...
Representation ....----'
Figure 1: Overview of the "many maps" model.
representation. A final step called "canonicalization" cleans up the representations of the
individual segments.
Figure 2 shows the effect of an M-P construction that breaks up CCC consonant clusters
by inserting a vowel after the first consonant, producing CiCCo The input in this case
is the Yawelmani word /?ugnhin/ "drinks", and the desired insertion is indicated in Pderiv. The mapping matrix derives the P-Ievel representation right-justified in the buffer,
with no segment gaps or collisions. It can do this even when mutliple simultaneous
insertions and deletions are being performed. But it cannot perform arbitrary sequence
manipulations, such as reversing all the segments of an utterance. Further details of the
matrix architecture are given in (Touretzlcy, 1989) and (Wheeler and Touretzky, 1989).
3 ITERATIVE PHENOMENA
Several types of phonological processes operate on groups of adjacent segments, often by
making them more similar to an immediately preceding (or following) trigger segment.
Vowel harmony and voicing assimilation are two examples. In Yawelmani, vowel harmony takes the following form: an [ahigh] vowel that is preceded by an [ahigh] round
vowel becomes round and back. In the form Ido:s+aV "might report", the non-round,
back vowel Ia! is [-high], as is the preceding round vowel/o/. Therefore the Ia! becomes
round, yielding the surface form [do:soIJ. Similarly, in Idub+hin/ "leads by the hand", the
[+highJ vowel Ii! is preceded by the [+high] round vowel lui, so the (II becomes round and
back, giving [dubhun]. In /bok'+hinl "finds", the Ii! does not undergo harmony because
it differs in height from the preceding vowel.
A Computational Basis for Phonology
P-Deriv:
mut
del
ins
-
-
-
-
-
-
i
h
-
-
-
-
1
-
-
n
g
u
?
M-Level:
-
-
+
n
-
-
n
1
h
h
n
n
P-Level:
i
i
g
u
M-PMapping
Matrix
?
Figure 2: Perfonning an insertion via the M-P mapping matrix.
Hannony is described as an iterative process because it can apply to entire sequences of
vowels, as in the following derivation:
It'ul+sit+hin/
It'ul+sut+hinl
It'ul+sut+hunl
"burns for"
harmony on second vowel
harmony on third vowel
In Yawelmani we saw an epenthesis process that inserts a high vowel Ii! to break up
lengthy consonant clusters. Epenthetic vowels may either undergo or block hannony.
With the word /logw+xa! "let's pulverize", epenthesis inserts an Ii! to break up the Igwx!
cluster, producing /logiw+xa!. Now the Ia! is preceded by a [+high, -round] vowel, so
hannony does not apply, whereas in Ido:s+al/, which has the same sequence of underlying
vowels, it did. This is an instance of epenthesis blocking hannony. In other environments
the epenthetic vowel may itself undergo hannony. For example:
/?ugn+hinl
/?uginhinl
/?ugunhin/
I?ugunhun/
"drinks"
epenthesis
harmony on epenthetic vowel
harmony on third vowel
The standard generative phonology analysis of hannony utilizes the following rule, applying after epenthesis, that is supposed to iterate through the utterance from left to right,
changing one vowel at a time:
375
376
Touretzky and Wheeler
+SYll ]
+round
]
[ +syll
[
[
1
high -- +back
/
:~;~d Co _
a
Lakoff offers an alternative account of epenthesis and harmony that eliminates iteration.
He states epenthesis as an M-P construction:
M:
C
P:
[]
I
C
i
{C.#}
I
[]
The harmony rule is stated as a P-Ievel well-formedness condition that applies simultaneously throughout the buffer:
P:
If [+syll. +round. ahigh] Co X.
then if X = [+syll. ahigh]. then X = [+round. +back].
Starting with /?ugn+hin/ at M-Ievel. Lakoff?s model would settle into a representation
of nugunhun/ at P-Ievel. We repeat again the crucial point that this representation is
not derived by sequential application of rules; it is merely licensed by one application
of epenthesis and two of harmony. The actual computation of the P-Ievel representation
would be performed by a parallel relaxation process. perhaps using simulated annealing.
that somehow determines the sequence that best satisfies all applicable constraints at
P-Ievel.
4 THE CLUSTERING MECHANISM
Our account of vowel harmony must differ from LakofCs because we do not wish to
rely on relaxation in our model. Instead. we introduce special clustering circuitry to
recognize sequences of segments that share certain properties. The clustering idea is
meant to be analogous to perceptual grouping in vision. Sequences of adjacent visuallysimilar objects are naturally perceived as a whole. A similar mechanism operating on
phonological sequences. although unprecedented in linguistic theory. does not appear
implausible. Crucial to our model is the principle that perceived sequences may be
operated on as a unit. This allows us to avoid iteration and give a fully-parallel account
of vowel harmony.
The clustering mechanism is controlled by a small number of language-specific parameters. The rule shown below is the P-F clustering rule for Yawelmani. Cluster type
[+syllabic] indicates that the rule looks only at vowels. (This is implemented by an
additional mapping matrix that extracts the vowel projection of the P-Ievel buffer. The
clustering mechanism actually looks at the output of this matrix rather than at the P-Ievel
buffer directly.) The trigger of a cluster is a round vowel of a given height. and the
elements are the subsequent adjacent vowels of matching height. Application of the rule
causes elements (but not triggers) to undergo a change; in this case. they become round
and back.
A Computational Basis for Phonology
Yawelmani vowel harmony - P-F mapping:
Cluster type: [+syllabic]
[+round, ahigh]
Trigger:
Element:
[ahigh]
Change:
[+round, +back]
The following hypothetical vowel sequence illustrates the application of this clustering
rule. Consonants are omitted for clarity:
I
234
5
6
7
8
9
1
U
i
e
0
0
a
i
+
+
trigger:
element:
i
+
+
+
+
The second vowel is round, so it's a trigger. Since the third and fourth vowels match it
in height, they become elements. The fifth vowel is [-highl, so it is not included in the
cluster. The sixth vowel triggers a new cluster because it's round; it is also [-high]. The
seventh and eighth vowels are also [-highl, so they can be elements, but the ninth vowel
is excluded from the cluster because is [+highl. Note that vowel 7 is an element, but
it also meets the specification for a trigger. Given a choice, our model prefers to mark
segments as elements rather than triggers because only elements undergo the specified
change. The distinction is moot in Yawlemani, where triggers are already round and
back, but it matters in other languages; see (Wheeler and Touretzky, 1989) for details.
Figures 2 and 3 together show the derivation of the Yawelmani word [?ugunhunl from
the underlying form /?ugn+hin/. In figure 2 an M-P construction inserted a high vowel.
In figure 3 the P-F clustering circuitry has examined the P-Ievel buffer and marked the
triggers and elements. Segments that were marked as elements then have the change
[+round, +backl written into their corresponding mutation slots in F-deriv. Finally, the
P-F mapping matrix produces the sequence /?ugunhun/ as the F-Ievel representation of
the utterance.
5 DISCUSSION
We could not justify the extra circuitry required for clustering if it were suitable only
for Yawelmani vowel harmony. The same mechanism handles a variety of other iterative
phenomena, including Slovak and Gidabal vowel shortening, Icelandic umlaut, and Russian voicing assimilation. The full mechanism has some additional parameters beyond
those covered in the discussion of Yawelmani. For example, clustering may proceed from
right-to-Ieft (as is the case in Russian) instead of from left-to-right Also, clusters may
be of either bounded or unbounded length. Bounded clusters are required for alternation
processes, such as Gidabal shortening. They cover exactly two segments: a trigger and
one element We are making a deliberate analogy here with metrical phonology (stress
systems), where unbounded feet may be of arbitrary length, but bounded feet always
contain exactly two syllables. No language has strictly trisyllabic feet We predict a similar constraint will hold for iterative phenomena when they are reformulated in parallel
clustering terms, i.e., no language requires bounded-length clusters with more than one
element
377
378
Touretzky and Wheeler
?
P-Ievel:
u
.
g
h
n
1
n
1
trigger
element
Clustering:
F-deriv:
.....-
n
u
n
I--
11
I-F-Ievel:
n
U
-~
u
h
..
n
u
g
u
u
T
"""--
J
.
P-FMapplng
Matrix
?
Figure 3: Clustering applied to Yawelmani vowel hannony.
A Computational Basis for Phonology
Our model makes many other predictions of constraints on human phonology, based
on limitations of the highly-structured "many maps" architecture. We are attempting to
verify these predictions, and also to extend the model to additional aspects of phonological
behavior, such as syllabification and stress.
Acknowledgements
This research was supported by a contract from Hughes Research Laboratories, by the
Office of Naval Research under contract number NOOOI4-86-K-0678, and by National
Science Foundation grant EET-8716324. We thank George Lakoff for encouragement
and support, John Goldsmith for helpful correspondence, and Gillette Elvgren III for
implementing the simulations.
References
Goldsmith, J. (to appear) Phonology as an intelligent system. To appear in a festschrift
for Leila Gleitman, edited by D. Napoli and J. Kegl.
Kenstowicz, M., and Kisseberth, C. (1979) Generative Phonology: Description and
Theory. San Diego, CA: Academic Press.
Lakoff, G. (1988) A suggestion for a linguistics with connectionist foundations. In D. S.
Touretzky, G. E. Hinton, and T. J. Sejnowski (eds.), Proceedings of the 1988 Connectionist
Models Summer School, pp. 301-314. San Mateo, CA: Morgan Kaufmann.
Lakoff, G. (1989) Cognitive phonology. Draft of paper presented at the UC-Berkeley
Workshop on Constraints vs Rules, May 1989.
Pinker, S., and Prince, A. (1988) On language and connectionism: analysis of a parallel
distributed processing model of language acquisition. In S. Pinker & J. Mehler (eds.),
Connections and Symbols. Cambridge, Massachusetts: MIT Press.
Rumelhart, D. E., and McClelland, J. L. (1986) On learning the past tenses of English
verbs. In J. L. McClelland and D. E. Rumelhart (eds.), Parallel Distributed Processing:
Explorations in the MicroStructJ&re of Cognition, volume 2. Cambridge, Massachusetts:
MIT Press.
Smolensky, P. (1986) Information processing in dynamical systems: foundations of harmony theory. In D. E. Rumelhart and J. L. McClelland (eds.), Parallel Distributed
Processing: Explorations in the MicroStructure of Cognition, volume 1. Cambridge,
Massachusetts: MIT Press.
Touretzky, D. S. (1989) Toward a connectionist phonology: the "many maps" approach to
sequence manipulation. Proceedings of the Eleventh Annual Conference of the Cognitive
Science Society, pp. 188-195. Hillsdale, NJ: Erlbaum.
Wheeler, D. W., and Touretzky, D. S. (1989) A connectionist implementation of cognitive
phonology. Technical report CMU-CS-89-144, Carnegie Mellon University, School of
Computer Science. To appear in G. Lakoff and L. Hyman (eds.), Proceedings of the UCBerkeley Phonology Workshop on Constraints vs. Rules. University of Chicago Press.
379
| 268 |@word simulation:1 past:1 yet:1 written:1 must:1 john:2 chicago:1 subsequent:1 v:2 alone:1 generative:2 nervous:1 draft:1 provides:1 height:4 unbounded:2 become:2 eleventh:1 introduce:2 inter:1 intricate:1 behavior:2 inspired:1 deirdre:1 actual:1 becomes:3 confused:1 underlying:4 bounded:4 sut:2 string:1 nj:1 berkeley:1 hypothetical:1 exactly:2 unit:1 grant:1 appear:6 producing:2 timing:1 critique:1 meet:1 might:1 burn:1 mateo:1 examined:1 co:2 directed:1 hughes:1 block:1 differs:1 wheeler:8 reject:1 projection:1 matching:1 word:4 suggest:1 cannot:1 applying:1 writing:1 map:6 primitive:1 straightforward:1 go:1 starting:1 immediately:1 rule:13 deriving:1 his:1 searching:1 handle:1 analogous:1 construction:8 trigger:13 diego:1 us:1 pa:2 element:14 rumelhart:4 blocking:1 inserted:1 edited:1 environment:1 slovak:1 insertion:5 touretzlcy:1 segment:11 purely:1 basis:6 easily:1 derivation:3 dialect:1 describe:1 sejnowski:1 quite:1 canonicalization:2 itself:1 final:1 sequence:16 unprecedented:1 inserting:1 mutliple:1 supposed:1 description:2 epenthetic:3 cluster:12 produce:2 hyman:1 object:1 develop:2 school:3 phonemic:1 implemented:3 c:1 involves:1 differ:1 foot:3 morphophonemic:1 exploration:2 human:7 settle:1 implementing:1 hillsdale:1 microstructure:1 formedness:5 connectionism:1 insert:2 strictly:1 hold:1 mapping:14 predict:1 cognition:2 circuitry:4 omitted:1 purpose:1 perceived:2 applicable:1 harmony:22 saw:1 icelandic:1 mit:3 always:1 rather:2 avoid:2 office:1 linguistic:3 derived:2 naval:1 indicates:1 helpful:1 entire:1 initially:1 development:1 proposes:1 constrained:2 special:1 uc:1 construct:1 phonological:8 never:1 represents:1 look:2 connectionist:8 others:2 report:2 intelligent:1 simultaneously:1 recognize:1 national:1 individual:1 mut:1 festschrift:1 vowel:45 highly:3 intra:1 yielding:1 operated:1 metrical:1 desired:1 prince:2 re:1 instance:1 modeling:1 cover:1 licensed:1 seventh:1 erlbaum:1 too:1 ido:2 contract:2 together:1 again:1 cognitive:4 american:1 account:5 matter:1 hannony:11 performed:3 break:3 pinker:3 parallel:8 mutation:3 phoneme:1 kaufmann:1 simultaneous:2 explain:1 implausible:1 touretzky:10 ed:5 lengthy:1 sixth:1 energy:2 acquisition:1 pp:2 naturally:1 noooi4:1 massachusetts:3 actually:1 back:8 feed:1 tum:1 specify:1 gillette:1 xa:2 hand:2 del:1 somehow:1 perhaps:1 indicated:1 russian:2 effect:1 lakoff:15 contain:1 verify:1 tense:1 excluded:2 read:2 laboratory:1 adjacent:3 round:19 stress:2 goldsmith:4 performs:2 reflection:1 hin:4 novel:2 common:1 specialized:1 preceded:3 overview:2 volume:2 extend:1 he:2 mellon:2 cambridge:3 encouragement:1 unconstrained:1 similarly:1 language:13 specification:1 surface:3 operating:1 reverse:1 manipulation:2 phonetic:2 certain:3 buffer:12 gleitman:1 alternation:1 morgan:1 minimum:1 additional:4 george:2 preceding:3 ii:5 multiple:1 full:1 technical:1 match:1 academic:1 offer:1 controlled:1 prediction:2 vision:1 cmu:1 iteration:2 proposal:1 justified:1 whereas:1 annealing:3 crucial:2 extra:1 eliminates:2 operate:1 virtually:1 undergo:5 intermediate:1 feedforward:1 easy:1 iii:1 iterate:1 variety:1 architecture:8 idea:2 ul:3 reformulated:1 proceed:1 cause:1 ieft:1 prefers:1 collision:1 clear:1 covered:1 ievel:23 shortening:2 mcclelland:4 deliberate:1 carnegie:2 discrete:1 write:1 group:1 drawn:1 changing:1 clarity:1 clean:1 relaxation:6 merely:1 fourth:1 throughout:1 utilizes:1 entirely:1 drink:2 summer:1 syllable:1 correspondence:1 annual:1 occur:1 constraint:7 phrased:1 syllabification:1 aspect:1 attempting:2 relatively:1 department:1 structured:1 combination:2 request:1 making:2 napoli:1 heart:1 mechanism:7 adopted:1 operation:4 apply:2 voicing:2 alternative:3 clustering:16 linguistics:2 phonology:18 giving:1 society:1 already:1 ccc:1 thank:1 simulated:3 reason:1 toward:1 length:3 relationship:1 difficult:1 troubling:1 stated:1 implementation:2 perform:1 av:1 kegl:1 hinton:1 ninth:1 verb:2 arbitrary:2 david:1 required:2 specified:1 connection:1 california:1 learned:1 deletion:3 distinction:1 beyond:1 below:1 dynamical:1 eighth:1 smolensky:2 genetically:1 including:1 ia:3 suitable:1 satisfaction:1 rely:1 scheme:3 extract:1 utterance:6 logw:1 acknowledgement:1 fully:1 interesting:1 limitation:2 suggestion:1 analogy:1 syllabic:2 foundation:3 principle:1 share:1 repeat:1 supported:1 english:1 highl:3 moot:1 perceptron:1 fifth:1 distributed:3 forward:1 san:2 epenthesis:8 eet:1 pittsburgh:3 consonant:4 search:2 iterative:6 why:2 nature:1 learn:2 ca:2 complex:2 did:1 whole:1 repeated:1 wellformedness:1 assimilation:3 wish:2 explicit:1 elvgren:1 perceptual:1 third:3 specific:1 symbol:1 appeal:1 deriv:6 derives:1 sit:1 grouping:1 workshop:2 sequential:1 illustrates:1 gap:1 desire:1 applies:1 cite:1 determines:1 satisfies:1 slot:2 goal:1 marked:2 change:7 included:1 determined:1 operates:1 reversing:1 lui:1 justify:1 called:3 m3:2 mark:1 support:1 meant:1 indian:1 perfonning:1 phenomenon:6 |
1,850 | 2,680 | New Criteria and a New Algorithm for Learning
in Multi-Agent Systems
Rob Powers
Computer Science Department
Stanford University
Stanford, CA 94305
[email protected]
Yoav Shoham
Computer Science Department
Stanford University
Stanford, CA 94305
[email protected]
Abstract
We propose a new set of criteria for learning algorithms in multi-agent
systems, one that is more stringent and (we argue) better justified than
previous proposed criteria. Our criteria, which apply most straightforwardly in repeated games with average rewards, consist of three requirements: (a) against a specified class of opponents (this class is a parameter
of the criterion) the algorithm yield a payoff that approaches the payoff
of the best response, (b) against other opponents the algorithm?s payoff
at least approach (and possibly exceed) the security level payoff (or maximin value), and (c) subject to these requirements, the algorithm achieve
a close to optimal payoff in self-play. We furthermore require that these
average payoffs be achieved quickly. We then present a novel algorithm,
and show that it meets these new criteria for a particular parameter class,
the class of stationary opponents. Finally, we show that the algorithm
is effective not only in theory, but also empirically. Using a recently
introduced comprehensive game theoretic test suite, we show that the
algorithm almost universally outperforms previous learning algorithms.
1
Introduction
There is rapidly growing interest in multi-agent systems, and in particular in learning algorithms for such systems. There is a growing body of algorithms proposed, and some arguments about their relative merits and domains of applicability (for example, [14] and [17]).
In [15] we survey much of this literature, and argue that it suffers from not having a clear
objective criteria with which to evaluate each algorithm (this shortcoming is not unique to
the relatively small computer science literature on multi-agent learning, and is shared by
the much vaster literature on learning in game theory). In [15] we also define five different
coherent agendas one could adopt, and identify one of them ? the agent-centric one ? as
particularly relevant from the computer science point of view.
In the agent-centric agenda one asks how an agent can learn optimally in the presence of
other independent agents, who may also be learning. To make the discussion precise we
will concentrate on algorithms for learning in known, fully observable two-player repeated
games, with average rewards. We start with the standard definition of a finite stage game
(aka normal form game):
Definition 1 A two-player stage game is a tuple G = (A1 , A2 , u1 , u2 ), where
? Ai is a finite set of actions available to player i
? ui : A1 ? A2 ? ? is a utility function for player i
Figure 1 shows two well-known games from the literature, to which we?ll refer again later.
In a repeated game the stage game is repeated, finitely or infinitely. The agent accumulates
rewards at each round; in the finite case the agent?s aggregate reward is the average of the
stage-game rewards, and in the infinite case it is the limit average (we ignore the subtlety
that arises when the limit does not exist, but this case does not present an essential problem).
While the vast majority of the literature on multi-agent learning (surprisingly) does not
start with a precise statement of objectives, there are some exceptions, and we discuss
them in the next section, including their shortcomings. In the following section we propose a stronger set of criteria that, we believe, does not suffer from these limitations. We
then present an algorithm that provably meets these stronger requirements. However, we
believe that all formal requirements ? including our own ? are merely baseline guarantees,
and any proposed algorithm must be subjected to empirical tests. While many previous
proposals provide empirical results, we think it is fair to say that our level of empirical
validation is unprecedented in the literature. We show the results of tests for all pairwise comparisons of major existing algorithms, using a recently-developed game theoretic
testbed called GAMUT [13] to systematically sample a very large space of games.
2
Previous criteria for multi-agent learning
To our knowledge, Bowling and Veloso [1] were the first in the AI community to explicitly
put forth formal requirements. Specifically they proposed two criteria:
Rationality: If the other players? policies converge to stationary policies then the learning
algorithm will converge to a stationary policy that is a best-response (in the stage game) to
the other players? policies.
Convergence: The learner will necessarily converge to a stationary policy.
Throughout this paper, we define a stationary policy as one that selects an action at each
point during the game by drawing from the same distribution, regardless of past history.
Bowling and Veloso considered known repeated games and proposed an algorithm that
provably meets their criteria in 2x2 games (games with two players and two actions per
player). Later, Conitzer and Sandholm [5] adopted the same criteria, and demonstrated an
algorithm meeting the criteria for all repeated games.
At first glance these criteria are reasonable, but a deeper look is less satisfying. First, note
that the property of convergence cannot be applied unconditionally, since one cannot ensure that a learning procedure converges against all possible opponents without sacrificing
rationality. So implicit in that requirement is some limitation on the class of opponents.
And indeed both [1] and [5] acknowledge this and choose to concentrate on the case of
self-play, that is, on opponents that are identical to the agent in question.
Dare
Y ield
Dare
0, 0
4, 1
Y ield
1, 4
2, 2
(a) Chicken
Cooperate
Def ect
Cooperate
3, 3
0, 4
Def ect
4, 0
1, 1
(b) Prisoner?s Dilemma
Figure 1: Example stage games. The payoff for the row player is given first in each cell,
with the payoff for the column player following.
We will have more to say about self-play later, but there are other aspects of these criteria
that bear discussion. While it is fine to consider opponents playing stationary policies,
there are other classes of opponents that might be as relevant or even more relevant; this
should be a degree of freedom in the definition of the problem. For instance, one might be
interested in the classes of opponents that can be modeled by finite automata with at most
k states; these include both stationary and non-stationary strategies.
We find the property of requiring convergence to a stationary strategy particularly hard to
justify. Consider the Prisoner?s Dilemma game in Figure 1. The Tit-for-Tat algorithm1
achieves an average payoff of 3 in self-play, while the unique Nash equilibrium of the
stage game has a payoff of only 1. Similarly, in the game of Chicken, also shown in
Figure 1, a strategy that alternates daring while its opponent yields and yielding while its
opponent dares achieves a higher expected payoff in self-play than any stationary policy
could guarantee. This problem is directly addressed in [2] and a counter-proposal made for
how to consider equilibria in repeated games. But there is also a fundamental issue with
these two criteria; they can both be thought of as a requirement on the play of the agent,
rather than the reward the agent receives.
Our final point regarding these two criteria is that they express properties that hold in the
limit, with no requirements whatsoever on the algorithm?s performance in any finite period.
But this question is not new to the AI community and has been addressed numerous times
in game theory, under the names of universal consistency, no-regret learning, and the Bayes
envelope, dating back to [9] (see [6] for an overview of this history). There is a fundamental
similarity in approach throughout, and we will take the two criteria proposed in [7] as being
representative.
Safety: The learning rule must guarantee at least the minimax payoff of the game. (The
minimax payoff is the maximum expected value a player can guarantee against any possible
opponent.)
Consistency: The learning rule must guarantee that it does at least as well as the best
response to the empirical distribution of play when playing against an opponent whose
play is governed by independent draws from any fixed distribution.
They then define universal consistency as the requirement that a learning rule do at least as
well as the best response to the empirical distribution regardless of the actual strategy the
opponent is employing (this implies both safety and consistency) and show that a modification of the fictitious play algorithm [3] achieves this requirement. A limitation common
to these game theory approaches is that they were designed for large-population games and
therefore ignore the effect of the agent?s play on the future play of the opponent. But this
can pose problems in smaller games. Consider the game of Prisoner?s Dilemma once again.
Even if the opponent is playing Tit-for-Tat, the only universally consistent strategy would
be to defect at every time step, ruling out the higher payoff achievable by cooperating.
3
A new set of criteria for learning
We will try to take the best of each proposal and create a joint set of criteria with the
potential to address some of the limitations mentioned above.
We wish to keep the notion of optimality against a specific set of opponents. But instead of
restricting this set in advance, we?ll make this a parameter of the properties. Acknowledging that we may encounter opponents outside our target set, we will also incorporate the
requirement of safety, which guarantees we achieve at least the security value, also known
1
The Tit-for-Tat algorithm cooperates in the first round and then for each successive round plays
the action its opponent played in the previous round.
as the maximin payoff, for the stage game. As a possible motivation for our approach,
consider the game of Rock-Paper-Scissors, which despite its simplicity has motivated several international tournaments. While the unique Nash equilibrium policy is to randomize,
the winners of the tournaments are those players who can most effectively exploit their
opponents who deviate without being exploited in turn.
The question remains of how best to handle self-play. One method would be to require that
a proposed algorithm be added to the set of opponents it is required to play a best response
to. While this may seem appealing at first glance, it can be a very weak requirement on the
actual payoff the agent receives. Since our opponent is no longer independent of our choice
of strategy, we can do better than settling for just any mutual best response, and try to maximize the value we achieve as well. We therefore propose requiring the algorithm achieve
at least the value of some Nash equilibrium that is Pareto efficient over the set of Nash
equilibria.2 Similarly, algorithms exist that satisfy ?universal consistency? and if played
by all agents will converge to a correlated equilibria[10], but this result provides an even
weaker constraint on the actual payoff received than convergence to a Nash equilibrium.
Let k be the number of outcomes for the game and b the maximum possible difference in
payoffs across the outcomes. We require that for any choice of ? > 0 and ? > 0 there exist
a T0 , polynomial in 1? , 1? , k, and b, such that for any number of rounds t > T0 the algorithm
achieves the following payoff guarantees with probability at least 1 ? ?.
Targeted Optimality: When the opponent is a member of the selected set of opponents,
the average payoff is at least VBR ??, where VBR is the expected value of the best response
in terms of average payoff against the actual opponent.
Compatibility: During self-play, the average payoff is at least Vself P lay ? ?, where
Vself P lay is defined as the minimum value achieved by the player in any Nash equilibrium
that is not Pareto dominated by another Nash equilibrium.
Safety: Against any opponent, the average payoff is at least Vsecurity ? ?, with Vsecurity
defined as max?1 ??1 min?2 ??2 EV (?1 , ?2 ).3
4
An algorithm
While we feel designing algorithms for use against more complex classes of opponent is
critical, as a minimal requirement we first show an algorithm that meets the above criteria
for the class of stationary opponents that has been the focus of much of the existing work.
Our method incorporates modifications of three simple strategies: Fictitious Play [3], Bully
[12], and the maximin strategy in order to create a more powerful hybrid algorithm.
Fictitious Play has been shown to converge in the limit to the best response against a stationary opponent. Each round it plays its best response to the most likely stationary opponent given the history of play. Our implementation uses a somewhat more generous
best-response calculation so as to achieve our performance requirements during self-play.
BR? (?) ? arg max(EOV (x, ?)), 4
x?X(?,?)
where X(?, ?) = {y ? ?1 : EV (y, ?) ? max (EV (z, ?)) ? ?}
z??1
2
An outcome is Pareto efficient over a set if there is no other outcome in that set with a payoff at
least as high for every agent and strictly higher for at least one agent.
3
Throughout the paper, we use EV (?1 , ?2 ) to indicate the expected payoff to a player for playing
strategy ?1 against an opponent playing ?2 and EOV (?1 , ?2 ) as the expected payoff the opponent
achieves. ?1 and ?2 are the sets of mixed strategies for the agent and its opponent respectively.
4
Note that BR0 (?) is a member of the standard set of best response strategies to ?.
We extend the Bully algorithm to consider the full set of mixed strategies and again maximize our opponent?s value when multiple strategies yield equal payoff for our agent.
BullyM ixed ? arg max(EOV (x, BR(x))),
x?X
where X = {y ? ?1 : EV (y, BR0 (y)) = max (EV (z, BR0 (z)))}
z??1
The maximin strategy is defined as
M aximin ? arg max min EV (?1 , ?2 )
?1 ??1
?2 ??2
We will now show how to combine these strategies into a single method satisfying all
three criteria. In the code shown below, t is the current round, AvgV aluem is the average value achieved by the agent during the last m rounds, VBully is shorthand for
EV (BullyM ixed, BR0 (BullyM ixed)), and dtt21 represents the distribution of opponent
actions for the period from round t1 to round t2 .
Set strategy = BullyMixed
for ?1 time steps
Play strategy
for ?2 time steps
if (strategy == BullyMixed AND AvgV alueH < VBully ? ?1 )
With probability, p, set strategy = BR?2 (dt0 )
Play strategy
if ||d?01 ? dtt??1 || < ?3
Set bestStrategy = BR?2 (dt0 )
else if (strategy == BullyMixed AND AvgV alueH > VBully ? ?1 )
Set bestStrategy = BullyMixed
else
Set bestStrategy = BestResponse
while not end of game
if avgV aluet??0 < Vsecurity ? ?0
Play maximin strategy for ?3 time steps
else
Play bestStrategy for ?3 time steps
The algorithm starts out with a coordination/exploration period in which it attempts to determine what class its opponent is in. At the end of this period it chooses one of three
strategies for the rest of the game. If it determines its opponent may be stationary it settles
on a best response to the history up until that point. Otherwise, if the BullyMixed strategy
has been performing well it maintains it. If neither of these conditions holds, it adopts a
default strategy, which we have set to be the BestResponse strategy. This strategy changes
each round, playing the best response to the maximum likelihood opponent strategy based
on the last H rounds of play. Once one of these strategies has been selected, the algorithm plays according to it whenever the average value meets or exceeds the security level,
reverting to the maximin strategy if the value drops too low.
Theorem 1 Our algorithm satisfies the three properties stated in section 3 for the class of
stationary opponents, with a T0 proportional to ( b? )3 1? .
This theorem can be proven for all three properties using a combination of basic probability
theory and repeated applications of the Hoeffding inequality [11], but the proof itself is
prohibitively long for inclusion in this publication.
5
Empirical results
Although satisfying the criteria we put forth is comforting, we feel this is but a first step
in making a compelling argument that an approach might be useful in practice. Traditionally, researchers suggesting a new algorithm also include an empirical comparison of the
algorithm to previous work. While we think this is a critical component of evaluating an
algorithm, most prior work has used tests against just one or two other algorithms on a
very narrow set of test environments, which often vary from researcher to researcher. This
practice has made it hard to consistently compare the performance of different algorithms.
In order to address this situation, we?ve started to code a collection of existing algorithms.
Combining this set of algorithms with a wide variety of repeated games from GAMUT
[13], a game theoretic test suite, we have the beginnings of a comprehensive testbed for
multi-agent learning algorithms. In the rest of this section, we?ll concentrate on the results
for our algorithm, but we hope that this testbed can form the foundation for a broad, consistent framework of empirical testing in multi-agent learning going forward. For all of
our environments we conducted our tests using a tournament format, where each algorithm
plays all other algorithms including itself.
0.500
MetaStrategy
StochFP
StochIGA
WoLF-PHC
BullyMixed
MiniMax
0.400
0.300
0.200
0.100
-M
ax
Jo
in
tQ
Hy
pe
r-Q
Lo
ca
lQ
St
oc
hF
P
St
oc
hI
GA
W
oL
FPH
C
m
et
aS
tra
te
gy
M
Ra
nd
o
Bu
lly
M
ini
M
ax
Bu
lly
M
i xe
d
-
Figure 2: Average value for last 20K rounds (of 200K) across all games in GAMUT.
Let us first consider the results of a tournament over a full set of games in GAMUT. Figure
2 portrays the average value achieved by each agent (y-axis) averaged over all games, when
playing different opponents (x-axis). The set of agents includes our strategy (MetaStrategy), six different adaptive learning approaches (Stochastic Fictitious Play [3,8], Stochastic
IGA[16], WoLF-PHC[1], Hyper-Q learning[18], Local Q-learning[19], and JointQ-Max[4]
(which learns Q-values over the joint action space but assumes its opponent will cooperate
to maximize its payoff)), and four fixed strategies (BullyMixed, Bully[12], the maximin
strategy, and Random (which selects a stationary mixed strategy at random)). We have
chosen a subset of the most successful algorithms to display on the graph. Against the four
stationary opponents, all of the adaptive learners fared equally well, while fixed strategy
players achieved poor rewards. In contrast, BullyMixed fared well against the adaptive
algorithms. As desired, our new algorithm combined the best of these characteristics to
achieve the highest value against all opponents except itself. It fares worse than BullyMixed
since it will always yield to BullyMixed, giving away the more advantageous outcome in
games like Chicken. However, when comparing how each agent performs in self-play,
our algorithm scores quite well, finishing a close second to Hyper-Q learning while the
two Bully algorithms finish near last. Hyper-Q is able to gain in self-play by occasionally
converging to outcomes with high social welfare that our strategy does not consider.
MetaStrategy
WoLF-PHC
StochFP
BullyMixed
StochIGA
MiniMax
100%
95%
90%
85%
80%
TwoByTwoGame
ShapleysGame
TravelersDilemma
Rochambeau
RandomZeroSum
RandomGame
RandomCompound
PrisonersDilemma
MinimumEffort
MajorityVoting
MatchingPennies
HawkAndDove
GrabTheDollar
GameWActionPar
DispersionGame
CovariantGame
CoordinationGame
Chicken
BertrandOligopoly
BattleOfTheSexes
70%
CollaborationGame
75%
Figure 3: Percent of maximum value for last 20K rounds (of 200K) averaged across all
opponents for selected games in GAMUT. The rewards were divided by the maximum
reward achieved by any agent to make visual comparisons easier.
So far we?ve seen that our new algorithm performs well when playing against a variety of
opponents. In Figure 3 we show the reward for each agent, averaged across the set of possible opponents for a selection of games in GAMUT. Once again our algorithm outperforms
the existing algorithms in nearly all games. When it fails to achieve the highest reward it
often appears to be due to its policy of ?generosity?; in games where it has multiple actions
yielding equal value, it chooses a best response that maximizes its opponent?s value.
The ability to study how individual strategies fare in each class of environment reflects an
advantage of our more comprehensive testing approach. In future work, this data can be
used both to aid in the selection of an appropriate algorithm for a new environment and to
pinpoint areas where an algorithm might be improved. Note that we use environment here
to indicate a combination of both the game and the distribution over opponents.
6
Conclusions and Future Work
Our objective in this work was to put forth a new set of criteria for evaluating the performance of multi-agent learning algorithms as well as propose a more comprehensive method
for empirical testing. In order to motivate this new approach for vetting algorithms, we have
presented a novel algorithm that meets our criteria and outperforms existing algorithms in a
wide variety of environments. We are continuing to work actively to extend our approach.
In particular, we wish to demonstrate the generality of our approach by providing algorithms that calculate best response to different sets of opponents (conditional strategies,
finite automata, etc.) Additionally, the criteria need to be generalized for n-player games
and we hope to combine our method for known games with methods for learning the structure of the game, ultimately devising new algorithms for unknown stochastic games.
Acknowledgements
This work was supported in part by a Benchmark Stanford Graduate Fellowship, DARPA
grant F30602-00-2-0598, and NSF grant IIS-0205633.
References
[1] Bowling, M. & Veloso, M. (2002). Multiagent learning using a variable learning rate. In Artificial
Intelligence, 136, pp. 215-250.
[2] Brafman, R. & Tennenholtz, M. (2002). Efficient Learning Equilibrium. In Advances in Neural
Information Processing Systems 15.
[3] Brown, G. (1951). Iterative Solution of Games by Fictitious Play. In Activity Analysis of Production and Allocation. New York: John Wiley and Sons.
[4] Claus, C. & Boutilier, C. (1998). The dynamics of reinforcement learning in cooperative multiagent systems. In Proceedings of the National Conference on Artificial Intelligence , pp. 746-752.
[5] Conitzer, V. & Sandholm, T. (2003). AWESOME: A General Multiagent Learning Algorithm that
Converges in Self-Play and Learns a Best Response Against Stationary Opponents. In Proceedings
of the 20th International Conference on Machine Learning, pp. 83-90, Washington, DC.
[6] Foster, D. & Vohra, R. (1999). Regret in the on-line decision problem. ?Games and Economic
Behavior? 29:7-36.
[7] Fudenberg, D. & Levine, D. (1995) Universal consistency and cautious fictitious play. Journal of
Economics Dynamics and Control 19:1065-1089.
[8] Fudenberg, D. & Levine, D. (1998). The theory of learning in games. MIT Press.
[9] Hannan, J. (1957) Approximation to Bayes risk in repeated plays. Contributions to the Theory of
Games 3:97-139.
[10] Hart, S. & Mas-Colell, A. (2000). A simple adaptive procedure leading to correlated equilibrium.
In Econometrica, Vol. 68, No. 5, pages 1127-1150.
[11] Hoeffding, W. (1956). On the distribution of the number of successes in independent trials.
Annals of Mathematical Statistics 27:713-721.
[12] Littman, M. & Stone, P. (2001). Implicit Negotiation in Repeated Games. In Proceedings of the
Eighth International Workshop on Agent Theories, Architectures, and Languages, pp. 393-404.
[13] Nudelman, E., Wortman, J., Leyton-Brown, K., & Shoham, Y. (2004). Run the GAMUT: A
Comprehensive Approach to Evaluating Game-Theoretic Algorithms. AAMAS-2004. To Appear.
[14] Sen, S. & Weiss, G. (1998). Learning in multiagent systems. In Multiagent systems: A modern
introduction to distributed artificial intelligence, chapter 6, pp. 259-298, MIT Press.
[15] Shoham, Y., Powers, R., & Grenager, T. (2003). Multi-Agent Reinforcement Learning: a critical
survey. Technical Report.
[16] Singh, S., Kearns, M., & Mansour, Y. (2000). Nash convergence of gradient dynamics in generalsum games. In Proceedings of UAI-2000, pp. 541-548, Morgan Kaufman.
[17] Stone, P. & Veloso, M. (2000). Multiagent systems: A survey from a machine learning perspective. Autonomous Robots, 8(3).
[18] Tesauro, G. (2004). Extending Q-Learning to General Adaptive Multi-Agent Systems. In Advances in Neural Information Processing Systems 16.
[19] Watkins, C. & Dayan, P. (1992). Technical note: Q-learning. Machine Learning, 8(3):279-292.
| 2680 |@word trial:1 polynomial:1 achievable:1 stronger:2 nd:1 advantageous:1 tat:3 asks:1 score:1 daring:1 outperforms:3 existing:5 past:1 current:1 comparing:1 must:3 john:1 designed:1 drop:1 stationary:18 intelligence:3 selected:3 devising:1 beginning:1 vbr:2 provides:1 successive:1 five:1 mathematical:1 ect:2 shorthand:1 combine:2 pairwise:1 expected:5 ra:1 indeed:1 behavior:1 growing:2 multi:11 ol:1 fared:2 actual:4 maximizes:1 what:1 kaufman:1 developed:1 whatsoever:1 suite:2 guarantee:7 every:2 prohibitively:1 comforting:1 control:1 grant:2 conitzer:2 appear:1 safety:4 t1:1 metastrategy:3 local:1 limit:4 despite:1 accumulates:1 meet:6 cooperates:1 might:4 tournament:4 graduate:1 averaged:3 unique:3 testing:3 practice:2 regret:2 procedure:2 area:1 empirical:9 universal:4 shoham:4 thought:1 cannot:2 close:2 ga:1 selection:2 put:3 risk:1 demonstrated:1 regardless:2 economics:1 automaton:2 survey:3 simplicity:1 rule:3 population:1 handle:1 notion:1 traditionally:1 autonomous:1 feel:2 annals:1 target:1 play:34 rationality:2 us:1 designing:1 satisfying:3 particularly:2 lay:2 dare:3 cooperative:1 levine:2 calculate:1 counter:1 highest:2 mentioned:1 environment:6 nash:8 ui:1 reward:11 econometrica:1 littman:1 dynamic:3 ultimately:1 motivate:1 singh:1 tit:3 dilemma:3 learner:2 joint:2 iga:1 darpa:1 chapter:1 effective:1 shortcoming:2 artificial:3 aggregate:1 hyper:3 outside:1 outcome:6 ixed:3 whose:1 dt0:2 stanford:7 quite:1 say:2 drawing:1 otherwise:1 ability:1 statistic:1 grenager:1 think:2 itself:3 final:1 advantage:1 unprecedented:1 rock:1 sen:1 propose:4 relevant:3 combining:1 rapidly:1 achieve:7 forth:3 cautious:1 convergence:5 requirement:14 extending:1 converges:2 pose:1 finitely:1 received:1 c:2 implies:1 indicate:2 concentrate:3 stochastic:3 exploration:1 stringent:1 settle:1 require:3 strictly:1 hold:2 considered:1 normal:1 welfare:1 equilibrium:11 major:1 achieves:5 adopt:1 a2:2 generous:1 vary:1 coordination:1 create:2 reflects:1 hope:2 mit:2 always:1 rather:1 publication:1 ax:2 focus:1 finishing:1 consistently:1 likelihood:1 aka:1 contrast:1 generosity:1 baseline:1 dayan:1 going:1 selects:2 interested:1 provably:2 compatibility:1 issue:1 arg:3 negotiation:1 mutual:1 equal:2 once:3 having:1 washington:1 identical:1 represents:1 broad:1 look:1 nearly:1 future:3 t2:1 report:1 modern:1 ve:2 comprehensive:5 individual:1 national:1 tq:1 attempt:1 freedom:1 interest:1 yielding:2 tuple:1 continuing:1 desired:1 sacrificing:1 minimal:1 instance:1 column:1 compelling:1 yoav:1 applicability:1 subset:1 successful:1 colell:1 conducted:1 wortman:1 too:1 optimally:1 straightforwardly:1 chooses:2 combined:1 st:2 fundamental:2 international:3 bu:2 quickly:1 jo:1 again:4 choose:1 possibly:1 hoeffding:2 worse:1 leading:1 actively:1 suggesting:1 potential:1 gy:1 includes:1 tra:1 satisfy:1 explicitly:1 scissors:1 later:3 view:1 try:2 start:3 bayes:2 maintains:1 hf:1 contribution:1 fph:1 acknowledging:1 who:3 characteristic:1 yield:4 identify:1 weak:1 vohra:1 researcher:3 history:4 suffers:1 whenever:1 definition:3 against:17 pp:6 proof:1 gain:1 knowledge:1 back:1 centric:2 appears:1 higher:3 response:16 improved:1 wei:1 generality:1 furthermore:1 just:2 stage:8 implicit:2 until:1 receives:2 glance:2 believe:2 name:1 effect:1 requiring:2 brown:2 round:14 ll:3 during:4 game:57 self:11 bowling:3 oc:2 criterion:26 generalized:1 ini:1 stone:2 theoretic:4 demonstrate:1 performs:2 percent:1 cooperate:3 br0:4 novel:2 recently:2 common:1 empirically:1 overview:1 winner:1 extend:2 fare:2 refer:1 ai:3 ield:2 consistency:6 similarly:2 inclusion:1 language:1 robot:1 similarity:1 longer:1 etc:1 own:1 perspective:1 tesauro:1 occasionally:1 inequality:1 success:1 meeting:1 xe:1 exploited:1 seen:1 minimum:1 morgan:1 somewhat:1 converge:5 maximize:3 period:4 determine:1 ii:1 full:2 multiple:2 hannan:1 exceeds:1 technical:2 bestresponse:2 veloso:4 calculation:1 long:1 divided:1 hart:1 equally:1 a1:2 converging:1 basic:1 lly:2 achieved:6 chicken:4 cell:1 justified:1 proposal:3 fellowship:1 fine:1 addressed:2 else:3 envelope:1 rest:2 claus:1 subject:1 member:2 incorporates:1 bully:4 seem:1 near:1 presence:1 exceed:1 variety:3 awesome:1 finish:1 architecture:1 economic:1 regarding:1 br:4 t0:3 motivated:1 six:1 utility:1 suffer:1 york:1 action:7 boutilier:1 useful:1 clear:1 exist:3 nsf:1 per:1 vol:1 express:1 four:2 neither:1 vast:1 graph:1 defect:1 merely:1 cooperating:1 vself:2 run:1 powerful:1 almost:1 throughout:3 reasonable:1 ruling:1 draw:1 decision:1 def:2 hi:1 played:2 display:1 activity:1 constraint:1 x2:1 hy:1 dominated:1 u1:1 aspect:1 argument:2 optimality:2 min:2 performing:1 relatively:1 format:1 department:2 according:1 alternate:1 combination:2 poor:1 sandholm:2 smaller:1 across:4 son:1 appealing:1 rob:1 modification:2 making:1 remains:1 discus:1 turn:1 reverting:1 merit:1 subjected:1 end:2 adopted:1 available:1 opponent:50 apply:1 away:1 appropriate:1 encounter:1 algorithm1:1 assumes:1 ensure:1 include:2 exploit:1 giving:1 f30602:1 objective:3 question:3 added:1 strategy:38 randomize:1 gradient:1 majority:1 argue:2 code:2 modeled:1 providing:1 statement:1 stated:1 agenda:2 implementation:1 policy:10 unknown:1 benchmark:1 finite:6 acknowledge:1 payoff:28 situation:1 precise:2 dc:1 mansour:1 community:2 introduced:1 required:1 specified:1 security:3 maximin:7 coherent:1 aximin:1 narrow:1 testbed:3 dtt:1 address:2 able:1 tennenholtz:1 below:1 ev:8 eighth:1 including:3 max:7 power:3 critical:3 settling:1 hybrid:1 generalsum:1 minimax:4 numerous:1 axis:2 gamut:7 started:1 unconditionally:1 dating:1 deviate:1 prior:1 literature:6 acknowledgement:1 relative:1 fully:1 multiagent:6 bear:1 mixed:3 limitation:4 proportional:1 fictitious:6 proven:1 allocation:1 validation:1 foundation:1 agent:34 degree:1 consistent:2 foster:1 systematically:1 playing:8 pareto:3 production:1 row:1 lo:1 surprisingly:1 last:5 supported:1 brafman:1 formal:2 weaker:1 deeper:1 wide:2 distributed:1 default:1 evaluating:3 adopts:1 made:2 collection:1 universally:2 forward:1 adaptive:5 reinforcement:2 employing:1 far:1 social:1 observable:1 ignore:2 keep:1 uai:1 iterative:1 additionally:1 learn:1 ca:3 necessarily:1 complex:1 domain:1 fudenberg:2 motivation:1 repeated:11 fair:1 aamas:1 body:1 representative:1 aid:1 wiley:1 fails:1 wish:2 lq:1 pinpoint:1 governed:1 pe:1 watkins:1 learns:2 theorem:2 specific:1 consist:1 essential:1 portrays:1 restricting:1 workshop:1 effectively:1 te:1 easier:1 likely:1 infinitely:1 visual:1 prisoner:3 subtlety:1 u2:1 wolf:3 leyton:1 determines:1 satisfies:1 ma:1 conditional:1 targeted:1 shared:1 hard:2 change:1 infinite:1 specifically:1 except:1 justify:1 kearns:1 called:1 player:16 exception:1 arises:1 incorporate:1 evaluate:1 phc:3 correlated:2 |
1,851 | 2,681 | Proximity graphs for clustering and manifold
learning
? Carreira-Perpin?
? an
Miguel A.
Richard S. Zemel
Dept. of Computer Science, University of Toronto
6 King?s College Road. Toronto, ON M5S 3H5, Canada
Email: {miguel,zemel}@cs.toronto.edu
Abstract
Many machine learning algorithms for clustering or dimensionality reduction take as input a cloud of points in Euclidean space, and construct
a graph with the input data points as vertices. This graph is then partitioned (clustering) or used to redefine metric information (dimensionality reduction). There has been much recent work on new methods for
graph-based clustering and dimensionality reduction, but not much on
constructing the graph itself. Graphs typically used include the fullyconnected graph, a local fixed-grid graph (for image segmentation) or a
nearest-neighbor graph. We suggest that the graph should adapt locally
to the structure of the data. This can be achieved by a graph ensemble
that combines multiple minimum spanning trees, each fit to a perturbed
version of the data set. We show that such a graph ensemble usually produces a better representation of the data manifold than standard methods;
and that it provides robustness to a subsequent clustering or dimensionality reduction algorithm based on the graph.
1
Introduction
Graph-based algorithms have long been popular, and have received even more attention
recently, for two of the fundamental problems in machine learning: clustering [1?4] and
manifold learning [5?8]. Relatively little attention has been paid to the properties and
construction methods for the graphs that these algorithms depend on.
A starting point for this study is the question of what constitutes a good graph. In the
applications considered here, the graphs are an intermediate form of representation, and
therefore their utility to some extent depends on the algorithms that they will ultimately be
used for. However, in the case of both clustering and manifold learning, the data points are
assumed to lie on some small number of manifolds. Intuitively, the graph should represent
these underlying manifolds well: it should avoid shortcuts that travel outside a manifold,
avoid gaps that erroneously disconnect regions of a manifold, and be dense within the
manifold and clusters. Also, while the algorithms differ with respect to connectedness, in
that clustering wants the graph to be disconnected, while for manifold learning the graph
should be connected, they both want at least the inside of the clusters, or dense areas of the
manifold, to be enhanced relative to the between-cluster, or sparse manifold connections.
Dataset
PSfrag replacements
MST
k-NNG
-ball graph
Delaunay
triangulation
Perturbed
dataset
Figure 1: Sensitivity to noise of proximity graphs. Top row: several proximity graphs
constructed on a noisy sample of points lying on a circle. Bottom row: the same graphs
constructed on a different sample; specifically, we added to each point Gaussian noise of
standard deviation equal to the length of the small segment shown in the centre of the
dataset (top row), built the graph, and drew it on the original dataset. This small perturbation can result in large changes in the graphs, such as disconnections, shortcuts or changes
in connection density.
Many methods employ simple graph constructions. A fully-connected graph is used for
example in spectral clustering and multidimensional scaling, while a fixed grid, with each
point connecting to some small fixed number of neighbors in a pre-defined grid of locations, is generally used in image segmentation. An -ball, in which each point connects
to points within some distance , and k-nearest neighbors (k-NNG) are generalization of
these approaches, as they take into account distance in some features associated with each
point instead of simply the grid locations. The -ball or k-NNG provide an improvement
over the fully-connected graph or fixed grid (clustering: [3, 9]; manifold learning: [5, 7]).
These traditional methods contain parameters (, k) that strongly depend on the data; they
generally require careful, costly tuning, as typically graphs must be constructed for a range
of parameter values, the clustering or dimensionality-reduction algorithm run on each, and
then performance curves compared to determine the best settings. Figure 1 shows that these
methods are quite sensitive to sparsity and noise in the data points, and that the parameters
should ideally vary within the data set. It also shows that other traditional graphs (e.g. the
Delaunay triangulation) are not good for manifolds, since they connect points nonlocally.
In this paper we propose a different method of graph construction, one based on minimum spanning trees (MSTs). Our method involves an ensemble of trees, each built on a
perturbed version of the data. We first discuss the motivation for this new type of graph,
and then examine its robustness properties, and its utility to both subsequent clustering or
dimensionality reduction methods.
2
Two new types of proximity graphs
A minimum spanning tree is a tree subgraph that contains all the vertices and has a minimum sum of edge weights. As a skeleton of a data set, the MST has some good properties:
it tends to avoid shortcuts between branches (typically caused by long edges, which are
contrary to the shortest-length criterion) and it gives a connected graph (usually a problem
for other methods with often-occurring random small groupings of points). In fact, the
MST was an early approach to clustering [10]. However, the MST is too sparse (having
only N ? 1 edges for an N -point data set, and no cycles) and is sensitive to noise. One
way to flesh it out and attain robustness to noise is to form an MST ensemble that combines
multiple MSTs; we give two different algorithms for this.
2.1
Perturbed MSTs (PMSTs)
Perturbed MSTs combine a number of MSTs, each fit to a perturbed version of the data
set. The perturbation is done through a local noise model that we estimate separately for
each data point based on its environment: point xi is perturbed by adding to it zero-mean
uniform noise of standard deviation si = rdi where di is the average distance to the k
nearest neighbors of xi , and r ? [0, 1]. In this paper we use k = 5 throughout and study the
effect of r. The locality of the noise model allows points to move more or less depending
on the local data structure around them and to connect to different numbers of neighbors at
different distances?in effect we achieve a variable k and .
To build the PMST ensemble, we generate T > 1 perturbed copies of the entire data set
according to the local noise model and fit an MST to each. The PMST ensemble assigns
a weight eij ? [0, 1] to the edge between points xi and xj equal to the average number
of times that edge appears on the trees. For T = 1 this gives the MST of the unperturbed
data set; for T ? ? it gives a stochastic graph where eij is the probability (in the Laplace
sense) of that edge under the noise model. The PMST ensemble contains at most T (N ? 1)
edges (usually much less). Although the algorithm is randomized, the PMST ensemble for
large T is essentially deterministic, and insensitive to noise by construction. In practice a
small T is enough; we use T = 20 in the experiments.
2.2
Disjoint MSTs (DMSTs)
Here we build a graph that is a deterministic collection of t MSTs that satisfies the property
that the nth tree (for n = 1, . . . , t) is the MST of the data subject to not using any edge
already in the previous 1, . . . , t ? 1 trees. One possible construction algorithm is an extension of Kruskal?s algorithm for the MST where we pick edges without replacement and
restart for every new tree. Specifically, we sort the list of N (N2?1) edges eij by increasing
distance dij and visit each available edge in turn, removing it if it merges two clusters (or
equivalently does not create a cycle); whenever we have removed N ? 1 edges, we go back
to the start of the list. We repeat the procedure t times in total.
The DMST ensemble consists of the union of all removed edges and contains t(N ? 1)
edges each of weight 1. The t parameter controls the overall density of the graph, which is
always connected; unlike or k (for the -ball or k-NNG), t is not a parameter that depends
locally on the data, and again points may connect to different numbers of neighbors at
different distances. We obtain the original MST for t = 1; values t = 2?4 (and usually
quite larger) work very well in practice. t need not be integer, i.e., we can fix the total
number of edges instead. In any case we should use t N2 .
2.3
Computational complexity
For a data set with N points, the computational complexity is approximately
O(T N 2 log N ) (PMSTs) or O(N 2 (log N + t)) (DMSTs). In both cases the resulting
graphs are sparse (number of edges is linear on number of points N ). If imposing an a
priori sparse structure (e.g. an 8-connected grid in image segmentation) the edge list is
much shorter, so the graph construction is faster. For the perturbed MST ensemble, the
perturbation of the data set results in a partially disordered edge list, which one should be
able to sort efficiently. The bottleneck in the graph construction itself is the computation of
pairwise distances, or equivalently of nearest neighbors, of a set of N points (which affects
the -ball and k-NNG graphs too): in 2D this is O(N log N ) thanks to properties of planar
geometry, but in higher dimensions the complexity quickly approaches O(N 2 ).
Overall, the real computational bottleneck is the graph postprocessing, typically O(N 3 ) in
spectral methods (for clustering or manifold learning). This can be sped up to O(cN 2 ) by
using sparsity (limiting a priori the edges allowed, thus approximating the true solution)
but then the graph construction is likewise sped up. Thus, even if our graphs are slightly
more costly to construct than the -ball or k-NNG, the computational savings are very large
if we avoid having to run the spectral technique multiple times in search for a good or k.
3
Experiments
We present two sets of experiments on the application of the graphs to clustering and manifold learning, respectively.
3.1
Clustering
In affinity-based clustering, our data is an N ? N affinity matrix W that defines a graph
(where nonzeros define edge weights) and we seek a partition of the graph that optimizes
a cost function, such as mincut [1] or normalized cut [2]. Typically the affinities are
wij = exp(? 12 (dij /?)2 ) (where dij is the problem-dependent distance between points
xi and xj ) and depend on a scale parameter ? ? (0, ?). This graph partitioning problem
is generally NP-complete, so approximations are necessary, such as spectral clustering algorithms [2]. In spectral clustering we seek to cluster in the leading eigenvectors of the
P
1
1
normalized affinity matrix N = D? 2 WD? 2 (where D = diag ( i wij ), and discarding
a constant eigenvector associated with eigenvalue 1). Spectral clustering succeeds only for
a range of values of ? where N displays the natural cluster structure of the data; if ? is
too small W is approximately diagonal and if ? is too large W is approximately a matrix
of ones. It is thus crucial to determine a good ?, which requires computing clusterings
over a range of ? values?an expensive computation since each eigenvector computation
is O(N 3 ) (or O(cN 2 ) under sparsity conditions).
Fig. 2 shows segmentation results for a grayscale image from [11] where the objective is
to segment the occluder from the underlying background, a hard task given the intensity
gradients. We use a standard weighted Euclidean distance on the data points (pixels) x =
(pixel location, intensity). One method uses the 8-connected grid (where each pixel is connected to its 8 neighboring pixels). The other method uses the PMST or DMST ensemble
(constrained to contain only edges in the 8-connected grid) under different values of the r, t
parameters; the graph has between 44% and 98% the number of edges in the 8-grid, depending on the parameter value. We define the affinity matrix as wij = eij exp(? 21 (dij /?)2 )
(where eij ? [0, 1] are the edge values). In both methods we apply the spectral clustering algorithm of [2]. The plot shows the clustering error (mismatched occluder area) for a
range of scales. The 8-connected grid succeeds in segmenting the occluder for ? ? [0.2, 1]
approximately, while the MST ensembles (for all parameter values tested) succeed for a
wider range?up to ? = ? in many cases. The reason for this success even for such high
? is that the graph lacks many edges around the occluder, so those affinities are zero no
matter how high the scale is. In other words, for clustering, our graphs enhance the inside
of clusters with respect to the bridges between clusters, and so ease the graph partitioning.
3.2
Manifold learning
For dimensionality reduction, we concentrate on applying Isomap [5], a popular and powerful algorithm. We first estimate the geodesic distances (i.e., along the manifold) g?ij be-
???? First 5 eigenvectors (except the constant one) ????
8-grid
?=?
PMST ensemble
8-grid
?2 = 1.6
PMST ensemble
8-grid
?1 = 0.5
PMST ensemble
Segmentation
60
1
2
PSfrag replacements
3
Original image
8-grid
graph
50
0
Clustering error
I
PMST ensemble
in 3D view
(x,y,intensity)
10
5
0
4
PMST, DMST
(various
parameters)
40
30
20
5
6
PSfrag replacements
y
10
8
7
6
8
4
0
x
10
2
0
?1
0
10
?
1
10
?
Figure 2: Using a proximity graph increases the scale range over which good segmentations
are possible. We consider segmenting the greyscale image at the bottom (an occluder over
a background) with spectral clustering, asking for K = 5 clusters. The color diagrams
represent the segmentation (column 1) and first 5 eigenvectors of the affinity matrix (except
the constant eigenvector, columns 2?4) obtained with spectral clustering, using a PMST
ensemble with r = 0.4 (upper row) or an 8-connectivity graph (lower row), for 3 different
scales: ?1 = 0.5, ?2 = 1.6 and ? = ?. The PMST ensemble succeeds at all scales
(note how several eigenvectors are constant over the occluder), while the 8-connectivity
graph progressively deteriorates as ? increases to give a partition of equal-sized clusters
for large scale. In the bottom part of the figure we show: the PMST ensemble graph in
3D space; the clustering error vs ? (where the right end is ? = ?) for the 8-connectivity
graph (thick blue line) and for various other PMST and DMST ensembles under various
parameters (thin lines). The PMST and DMST ensembles robustly (for many settings of
their parameters) give an almost perfect segmentation over a large range of scales.
tween pairs of points in the data set as the shortest-path lengths in a graph learned from the
data. Then we apply multidimensional scaling to these distances to obtain a collection of
low-dimensional points {yi }N
i=1 that optimally preserves the estimated geodesic distances.
In fig. 3 we show the results of applying Isomap using different graphs to two data sets
(ellipse and Swiss roll) for which we know the true geodesic distances gij . In a real application, since the true geodesic distances are unknown, error and variance cannot be
computed; an estimated residual variance has been proposed [5] to determine the optimal
graph parameter. For the perturbed MST ensemble, we binarize the edge values by making
1 any eij > 0. (It is often possible to refine the graph by zeroing edges with small eij ,
since this removes shortcuts that may have arisen by chance, particularly if T is large; but
it is difficult to estimate the right threshold reliably.) The plots show 3 curves as a function
of the graph parameter: the average error E in the geodesic distances; Isomap?s estimated
residual variance V? ; and the true residual variance V . From the plots we can see that V?
correlates well with V (though it underestimates it) and also with E for the Swiss roll, but
not for the ellipse; this can make the optimal graph parameter difficult to determine in a
real application. Given this, the fact that our graphs work well over a larger region of their
parameter space than the -ball or k-NNG graphs makes them particularly attractive.
The plots for the Swiss roll show that, while for the low noise case the -ball or k-NNG
graphs work well over a reasonable region of their parameter space, for the high noise case
this region decreases a lot, almost vanishing for the -ball. This is because for low values
of the parameter the graph is disconnected, while for high values it has multiple shortcuts;
the difficulty of the task is compounded by the small number of points used, N = 500 (an
unavoidable fact in high dimensions). However, for the PMSTs the region remains quite
wide and for the DMSTs the approximate region t ? [2, 8] gives good results. For very low
r or t = 1 the graph is the single MST, thus the large errors.
It is also important to realize that the range of the r parameter of the PMST ensemble does
not depend on the data, while the range for and k does. The range of the t parameter of
the DMST ensemble does depend on the data, but we have found empirically that t = 2?4
gives very good results with all data sets we have tried.
4
Discussion
One main contribution of this paper is to highlight the relatively understudied problem of
converting a data set into a graph, which forms an intermediate representation for many
clustering and manifold learning algorithms. A second contribution is novel construction
algorithms, which are: easy to implement, not expensive to compute, robust across many
noise levels and parameter settings, and useful for clustering and manifold learning. In
general, a careful selection of the graph construction algorithm makes the results of these
machine learning methods robust, and avoids or limits the required parameter search. Finally, the combination of many graphs, formed from perturbed versions of the data, into an
ensemble of graphs, is a novel approach to the construction problem.
Our idea of MST ensembles is an extension to graphs of the well-known technique of combining predictors by averaging (regression) or voting (classification), as is the regularizing
effect of training with noise [12]. An ensemble of predictors improves the generalization to
unseen data if the individual predictors are independent from each other and disagree with
each other, and can be explained by the bias-variance tradeoff. Unlike regression or classification, unsupervised graph learning lacks at present an error function, so it seems difficult
to apply the bias-variance framework here. However, we have conducted a wide range of
empirical tests to understand the properties of the ensemble MSTs, and to compare them
to the other graph construction methods, in terms of the error in the geodesic distances (if
known a priori). In summary, we have found that the variance of the error for the geodesic
Ellipse, high noise
Swiss roll, low noise
Swiss roll, high noise
0.8
50
Data set
0.6
0.4
50
0
0
?50
?50
10
10
0.2
5
PSfrag replacements
kr
t
0
PSfrag replacements
kr
t
?0.2
?0.4
?0.6
?0.8
?1.5
?1
?0.5
0
0.5
1
1.5
1.0332
1
5
0
?5
?10
?15
?10
?15
5
0
?5
PSfrag replacements
kr
t
10
-ball
0.75
1
0.5166
37.4454
0.5
PSfrag replacements
0.25
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1.0332
1
0
kr
t
1
k-NNG
0.7749
0.5166
0.5
0.2583
PSfrag replacements
r
t
k
0.25
0
5
10
15
20
0
30
25
PMST ensemble
DMST ensemble
15
1
avg error in gd
resvar (estimated)
resvar (true)
24.9636
0.5
12.4818
0.25
0
1
2
3
4
5
6
7
8
1
0.7749
37.4454
0.75
24.9636
0.5
0.5
0.2583
0
PSfrag replacements
k
t
r
0.25
0
0.1
0.2
0.3
0.4
0.5
0.6
37.4454
12.4818
0.25
12.4818
0
0
5
10
15
20
10
20
25
2
3
4
5
6
7
8
1
37.4454
0.75
0.5
24.9636
0.5
0.25
12.4818
0.25
0
5
10
15
20
0.75
0
0.1
0.2
0.3
r
0
30
0.4
0.5
0.6
37.4454
1
37.4454
0.75
0.5
24.9636
0.5
0.25
12.4818
0.25
0
0.7
0
1
49.9272
0
0.1
0.2
0.3
r
0.4
0.5
0.75
5
10
15
t
20
25
0.6
V,
V?
0
0.7
1
avg error in gd
resvar (estimated)
resvar (true)
37.4454
0.75
0.5
24.9636
0.5
0.25
12.4818
0.25
PSfrag replacements
kr
12.4818
0
0
30
25
49.9272
avg error in gd
resvar (estimated)
resvar (true)
24.9636
V,
V?
avg error in gd
resvar (estimated)
resvar (true)
PSfrag replacements
k
t
12.4818
0
10
avg error in gd
resvar (estimated)
resvar (true)
1
24.9636
9
V,
V?
k
37.4454
49.9272
PSfrag replacements
kr
t
15
49.9272
0
30
25
49.9272
1
0.25
5
1
1
k
0
0.5
0.2583
0
0.75
24.9636
0
0.7
0.75
0.5166
0
10
PSfrag replacements
r
t
avg error in gd
resvar (estimated)
resvar (true)
0.7749
kr
t
avg error in gd
resvar (estimated)
resvar (true)
0.75
0.5166
9
49.9272
avg error in gd
resvar (estimated)
resvar (true)
1.0332
E
PSfrag replacements
kr
10
5
0
?5
avg error in gd
resvar (estimated)
resvar (true)
0.75
1.0332
E
PSfrag replacements
k
t
?10
?15
49.9272
0.75
avg error in gd
resvar (estimated)
resvar (true)
E
PSfrag replacements
r
t
?15
PSfrag replacements
0.2583
kr
t
?10
avg error in gd
resvar (estimated)
resvar (true)
0.7749
E
PSfrag replacements
?5
15
49.9272
avg error in gd
resvar (estimated)
resvar (true)
0
0
30
0
5
10
15
20
25
V,
V?
0
30
t
Figure 3: Performance of Isomap with different graphs in 3 data sets: ellipse with N = 100
points, high noise; Swiss roll with N = 500 points, low and high noise (where high noise
means Gaussian with standard deviation equal to 9% of the separation between branches).
All plots show on the X axis the graph parameter (, k, r or t); on the left Y axis the
PN
average error in the geodesic distances (red curve, E = N12 i,j=1 |?
gij ? gij |); and on the
? DY ))
right Y axis Isomap?s estimated residual variance (solid blue curve, V? = 1 ? R2 (G,
2
?
and true residual variance (dashed blue curve, V = 1 ? R (G, DY )), where G and G
are the matrices of estimated and true geodesic distances, respectively, DY is the matrix of
Euclidean distances in the low-dimensional embedding, and R(A, B) is the standard linear
correlation coefficient, taken over all entries of matrices A and B. Where the curves for
-ball and k-NNG are missing, the graph was disconnected.
distances decreases for the ensemble when the individual graphs are sparse (e.g. MSTs as
used here, or -ball and k-NNG with low or k); but not necessarily when the graphs are
not sparse.
The typical cut [9, 13] is a clustering criterion that is based on the probability pij that
points xi and xj are in the same cluster over all possible partitions (under the Boltzmann
distribution for the mincut cost function). The pij need to be estimated: [9] use SwendsenWang sampling, while [13] use randomized trees sampling. However, these trees are not
used to define a proximity graph, unlike in our work.
An important direction for future work concerns the noise model for PMSTs. The model
we propose is isotropic, in that every direction of perturbation is equally likely. A better
way is to perturb points more strongly in directions likely to lie within the manifold and less
strongly in directions away from the manifold, using a method such as k nearest neighbors
to estimate appropriate directions. Preliminary experiments with such a manifold-aligned
model are very promising, particularly when the data is very noisy or its distribution on the
manifold is not uniform. The noise model can also be extended to deal with non-Euclidean
data by directly perturbing the similarities.
Acknowledgements
Funding provided by a CIHR New Emerging Teams grant.
References
[1] Zhenyu Wu and Richard Leahy. An optimal graph theoretic approach to data clustering: Theory
and its application to image segmentation. IEEE Trans. on Pattern Anal. and Machine Intel.,
15(11):1101?1113, November 1993.
[2] Jianbo Shi and Jitendra Malik. Normalized cuts and image segmentation. IEEE Trans. on
Pattern Anal. and Machine Intel., 22(8):888?905, August 2000.
[3] Pedro F. Felzenszwalb and Daniel P. Huttenlocher. Efficient graph-based image segmentation.
Int. J. Computer Vision, 59(2):167?181, September 2004.
[4] Romer Rosales, Kannan Achan, and Brendan Frey. Learning to cluster using local neighborhood structure. In ICML, 2004.
[5] Joshua B. Tenenbaum, Vin de Silva, and John C. Langford. A global geometric framework for
nonlinear dimensionality reduction. Science, 290(5500):2319?2323, December 22 2000.
[6] Sam T. Roweis and Lawrence K. Saul. Nonlinear dimensionality reduction by locally linear
embedding. Science, 290(5500):2323?2326, December 22 2000.
[7] Mikhail Belkin and Partha Niyogi. Laplacian eigenmaps for dimensionality reduction and data
representation. Neural Computation, 15(6):1373?1396, June 2003.
[8] Kilian Q. Weinberger and Lawrence K. Saul. Unsupervised learning of image manifolds by
semidefinite programming. In CVPR, 2004.
[9] Marcelo Blatt, Shai Wiseman, and Eytan Domany. Data clustering using a model granular
magnet. Neural Computation, 9(8):1805?1842, November 1997.
[10] C. T. Zahn. Graph-theoretical methods for detecting and describing gestalt clusters. IEEE
Trans. Computers, C?20(1):68?86, April 1971.
[11] Chakra Chennubhotla and Allan Jepson. EigenCuts: Half-lives of EigenFlows for spectral
clustering. In NIPS, 2003.
[12] Christopher M. Bishop. Neural Networks for Pattern Recognition. Oxford University Press,
New York, Oxford, 1995.
[13] Yoram Gdalyahu, Daphna Weinshall, and Michael Werman. Self organization in vision: Stochastic clustering for image segmentation, perceptual grouping, and image database organization. IEEE Trans. on Pattern Anal. and Machine Intel., 23(10):1053?1074, October 2001.
| 2681 |@word version:4 seems:1 seek:2 perpin:1 tried:1 paid:1 pick:1 solid:1 reduction:10 contains:3 nonlocally:1 daniel:1 wd:1 si:1 must:1 john:1 realize:1 mst:24 subsequent:2 partition:3 remove:1 plot:5 progressively:1 v:1 half:1 isotropic:1 vanishing:1 provides:1 detecting:1 toronto:3 location:3 along:1 constructed:3 psfrag:18 consists:1 combine:3 fullyconnected:1 redefine:1 inside:2 pairwise:1 allan:1 examine:1 occluder:6 little:1 increasing:1 provided:1 underlying:2 what:1 weinshall:1 eigenvector:3 emerging:1 every:2 multidimensional:2 voting:1 jianbo:1 control:1 partitioning:2 grant:1 segmenting:2 local:5 frey:1 tends:1 limit:1 oxford:2 path:1 approximately:4 connectedness:1 ease:1 range:11 gdalyahu:1 practice:2 union:1 implement:1 swiss:6 procedure:1 area:2 empirical:1 attain:1 pre:1 road:1 word:1 suggest:1 cannot:1 selection:1 applying:2 deterministic:2 missing:1 shi:1 go:1 attention:2 starting:1 assigns:1 leahy:1 embedding:2 n12:1 laplace:1 limiting:1 construction:12 enhanced:1 zhenyu:1 programming:1 us:2 expensive:2 particularly:3 recognition:1 cut:3 huttenlocher:1 database:1 bottom:3 cloud:1 region:6 connected:10 cycle:2 kilian:1 decrease:2 removed:2 environment:1 complexity:3 skeleton:1 ideally:1 geodesic:9 ultimately:1 depend:5 segment:2 various:3 zemel:2 outside:1 neighborhood:1 quite:3 larger:2 cvpr:1 tested:1 niyogi:1 unseen:1 itself:2 noisy:2 eigenvalue:1 propose:2 cihr:1 neighboring:1 aligned:1 combining:1 subgraph:1 achieve:1 roweis:1 cluster:13 produce:1 perfect:1 wider:1 depending:2 miguel:2 ij:1 nearest:5 received:1 c:1 involves:1 flesh:1 rosales:1 eigenflows:1 differ:1 concentrate:1 nng:11 thick:1 direction:5 stochastic:2 disordered:1 require:1 fix:1 generalization:2 preliminary:1 extension:2 proximity:6 lying:1 considered:1 around:2 exp:2 lawrence:2 werman:1 kruskal:1 vary:1 early:1 travel:1 sensitive:2 bridge:1 create:1 weighted:1 gaussian:2 always:1 avoid:4 pn:1 june:1 improvement:1 brendan:1 sense:1 dependent:1 typically:5 entire:1 wij:3 pixel:4 overall:2 classification:2 priori:3 constrained:1 equal:4 construct:2 saving:1 having:2 sampling:2 unsupervised:2 constitutes:1 thin:1 icml:1 future:1 np:1 richard:2 employ:1 belkin:1 preserve:1 individual:2 zahn:1 geometry:1 connects:1 replacement:19 organization:2 semidefinite:1 edge:26 necessary:1 shorter:1 tree:11 euclidean:4 circle:1 theoretical:1 column:2 asking:1 wiseman:1 cost:2 vertex:2 deviation:3 rdi:1 entry:1 uniform:2 predictor:3 dij:4 conducted:1 eigenmaps:1 too:4 optimally:1 connect:3 perturbed:11 gd:12 thanks:1 density:2 fundamental:1 sensitivity:1 randomized:2 enhance:1 connecting:1 quickly:1 michael:1 connectivity:3 again:1 unavoidable:1 leading:1 account:1 de:1 disconnect:1 matter:1 coefficient:1 jitendra:1 int:1 caused:1 depends:2 view:1 lot:1 red:1 start:1 sort:2 vin:1 shai:1 blatt:1 partha:1 contribution:2 formed:1 marcelo:1 roll:6 variance:9 efficiently:1 ensemble:31 likewise:1 m5s:1 whenever:1 email:1 underestimate:1 associated:2 di:1 dataset:4 popular:2 color:1 dimensionality:10 improves:1 segmentation:12 back:1 chakra:1 appears:1 higher:1 planar:1 april:1 done:1 though:1 strongly:3 correlation:1 langford:1 christopher:1 nonlinear:2 lack:2 defines:1 effect:3 contain:2 true:18 isomap:5 normalized:3 deal:1 attractive:1 self:1 criterion:2 complete:1 theoretic:1 silva:1 postprocessing:1 image:12 novel:2 recently:1 funding:1 sped:2 empirically:1 perturbing:1 insensitive:1 imposing:1 tuning:1 grid:14 zeroing:1 centre:1 similarity:1 delaunay:2 chennubhotla:1 recent:1 triangulation:2 optimizes:1 success:1 life:1 yi:1 joshua:1 minimum:4 converting:1 determine:4 shortest:2 dashed:1 branch:2 multiple:4 nonzeros:1 compounded:1 faster:1 adapt:1 long:2 dept:1 equally:1 visit:1 laplacian:1 regression:2 essentially:1 metric:1 vision:2 represent:2 arisen:1 achieved:1 background:2 want:2 separately:1 diagram:1 crucial:1 unlike:3 subject:1 december:2 contrary:1 integer:1 intermediate:2 enough:1 easy:1 xj:3 fit:3 affect:1 idea:1 cn:2 domany:1 tradeoff:1 bottleneck:2 utility:2 york:1 generally:3 useful:1 eigenvectors:4 locally:3 tenenbaum:1 generate:1 deteriorates:1 disjoint:1 estimated:18 blue:3 threshold:1 achan:1 graph:86 sum:1 run:2 powerful:1 throughout:1 almost:2 reasonable:1 wu:1 separation:1 dy:3 scaling:2 display:1 refine:1 erroneously:1 relatively:2 according:1 ball:12 disconnected:3 combination:1 across:1 slightly:1 sam:1 partitioned:1 making:1 intuitively:1 explained:1 taken:1 remains:1 discus:1 turn:1 describing:1 know:1 end:1 available:1 apply:3 away:1 spectral:10 appropriate:1 romer:1 robustly:1 robustness:3 weinberger:1 original:3 top:2 clustering:35 include:1 mincut:2 daphna:1 yoram:1 perturb:1 build:2 ellipse:4 approximating:1 move:1 objective:1 question:1 added:1 already:1 malik:1 costly:2 traditional:2 diagonal:1 september:1 affinity:7 gradient:1 distance:20 restart:1 manifold:25 extent:1 binarize:1 spanning:3 reason:1 kannan:1 length:3 equivalently:2 difficult:3 october:1 greyscale:1 anal:3 reliably:1 boltzmann:1 unknown:1 upper:1 disagree:1 november:2 extended:1 team:1 perturbation:4 august:1 canada:1 intensity:3 pair:1 required:1 connection:2 merges:1 learned:1 nip:1 trans:4 able:1 usually:4 pattern:4 sparsity:3 built:2 natural:1 difficulty:1 residual:5 nth:1 axis:3 geometric:1 acknowledgement:1 relative:1 fully:2 highlight:1 granular:1 pij:2 row:5 summary:1 repeat:1 copy:1 bias:2 disconnection:1 understand:1 mismatched:1 neighbor:8 wide:2 felzenszwalb:1 saul:2 mikhail:1 sparse:6 curve:6 dimension:2 avoids:1 collection:2 avg:12 correlate:1 gestalt:1 approximate:1 global:1 assumed:1 xi:5 grayscale:1 search:2 promising:1 robust:2 necessarily:1 constructing:1 diag:1 jepson:1 tween:1 dense:2 main:1 motivation:1 noise:23 n2:2 allowed:1 fig:2 intel:3 lie:2 perceptual:1 removing:1 discarding:1 bishop:1 unperturbed:1 list:4 r2:1 concern:1 grouping:2 adding:1 drew:1 kr:9 occurring:1 gap:1 locality:1 simply:1 eij:7 likely:2 partially:1 pedro:1 satisfies:1 chance:1 succeed:1 sized:1 king:1 careful:2 shortcut:5 change:2 hard:1 carreira:1 specifically:2 except:2 typical:1 averaging:1 total:2 gij:3 eytan:1 succeeds:3 college:1 h5:1 magnet:1 regularizing:1 |
1,852 | 2,682 | Optimal sub-graphical models
Mukund Narasimhan? and Jeff Bilmes?
Dept. of Electrical Engineering
University of Washington
Seattle, WA 98195
{mukundn,bilmes}@ee.washington.edu
Abstract
We investigate the problem of reducing the complexity of a graphical
model (G, PG ) by finding a subgraph H of G, chosen from a class of
subgraphs H, such that H is optimal with respect to KL-divergence. We
do this by first defining a decomposition tree representation for G, which
is closely related to the junction-tree representation for G. We then give
an algorithm which uses this representation to compute the optimal H ?
H. Gavril [2] and Tarjan [3] have used graph separation properties to
solve several combinatorial optimization problems when the size of the
minimal separators in the graph is bounded. We present an extension of
this technique which applies to some important choices of H even when
the size of the minimal separators of G are arbitrarily large. In particular,
this applies to problems such as finding an optimal subgraphical model
over a (k ? 1)-tree of a graphical model over a k-tree (for arbitrary k)
and selecting an optimal subgraphical model with (a constant) d fewer
edges with respect to KL-divergence can be solved in time polynomial in
|V (G)| using this formulation.
1
Introduction and Preliminaries
The complexity of inference in graphical models is typically exponential in some parameter of the graph, such as the size of the largest clique. Therefore, it is often required to find
a subgraphical model that has lower complexity (smaller clique size) without introducing
a large error in inference results. The KL-divergence between the original probability distribution and the probability distribution on the simplified graphical model is often used to
measure the impact on inference. Existing techniques for reducing the complexity of graphical models including annihilation and edge-removal [4] are greedy in nature and cannot
make any guarantees regarding the optimality of the solution. This problem is NP-complete
[9] and so, in general, one cannot expect a polynomial time algorithm to find the optimal
solution. However, we show that when we restrict the problem to some sets of subgraphs,
the optimal solution can be found quite quickly using a dynamic programming algorithm
in time polynomial in the tree-width of the graph.
1.1
Notation and Terminology
A graph G = (V, E) is said to be triangulated if every cycle of length greater than 3
has a chord. A clique of G is a non-empty set S ? V such that {a, b} ? E for all
?
This work was supported by NSF grant IIS-0093430 and an Intel Corporation Grant.
{b, c, d}
d
{c, f, g}
{b, c}
{b, e, c}
b
c
{f, c}
{c, e}
{e, c, f }
g
{b, e}
a
e
f
{a, b, e}
Figure 1: A triangulated graph G and a junction-tree for G
a, b ? S. A clique S is maximal if S is not properly contained in another clique. If
? and ? are non-adjacent vertices of G then a set of vertices S ? V \ {?, ?} is called
an (?, ?)-separator if ? and ? are in distinct components of G[V \ S]. S is a minimal
(?, ?)-separator if no proper subset of S is an (?, ?)-separator. S is said to be a minimal
separator if S is a minimal (?, ?)-separator for some non adjacent a, b ? V . If T =
(K, S) is a junction-tree for G (see [7]), then the nodes K of T correspond to the maximalcliques of G, while the links S correspond to minimal separators of G (We reserve the
terms vertices/edges for elements of G, and nodes/links for the elements of T ). If G is
triangulated, then the number of maximal cliques is at most |V |. For example, in the graph
G shown in Figure 1, K = {{b, c, d} , {a, b, e} , {b, e, c} , {e, c, f } , {c, f, g}}. The links
S of T correspond to minimal-separators of G in the following way. If Vi Vj ? S (where
Vi , Vj ? K and hence are cliques of G), then Vi ? Vj 6= ?. We label each edge Vi Vj ? S
with the set Vij = Vi ? Vj , which is a non-empty complete separator in G. The removal of
any link Vi Vj ? S disconnects T into two subtrees which we denote T (i) and T (j) (chosen
so that T (i) contains Vi ). We will let K(i) be the nodes of T (i) , and V (i) = ?V ?K (i) V be
the set of vertices corresponding to the subtree T (i) . The junction tree property ensures that
V (i) ? V (j) = Vi ? Vj = Vij . We will let G(i) be the subgraph induced by V (i) .
A graphical model is a pair (G, P ) where P is the joint probability distribution for random
variables X1 , X2 , . . . , Xn , and G is a graph with vertex set V (G) = {X1 , X2 , . . . , Xn }
such that the separators in G imply conditional independencies in P (so P factors according
to G). If G is triangulated, then the junction-tree algorithm can be used for exact inference
in the probability distribution P . The complexity of this algorithm grows with the treewidth
of G (which is one less than the size of the largest clique in G when G is triangulated). The
growth is exponential when P is a discrete probability distribution, thus rendering exact
inference for graphs with large treewidth impractical. Therefore, we seek another graphical
model (H, PH ) which allows tractable inference (so H should have lower treewidth than
G has). The general problem of finding a graphical model of tree-width at most k so as
to minimize the KL-divergence from a specified probability distribution is NP complete
for general k ([9]) However, it is known that this problem is solvable in polynomial time
(in |V (G)|) for some special cases cases (such as when G has bounded treewidth or when
k = 1 [1]). If (G, PG ) and (H, PH ) are graphical models, then we say that (H, PH ) is a
subgraphical model of (G, PG ) if H is a spanning subgraph of G. Note in particular that
separators in G are separators in H, and hence (G, PH ) is also a graphical model.
2
Graph Decompositions and Divide-and-Conquer Algorithms
For the remainder of the paper, we will be assuming that G = (V, E) is some triangulated
graph, with junction tree T = (K, S). As observed above, if Vi Vj ? S, then the removal
{b, c, d}
d
{b, c}
{b, e, c}
b
c
c
{c, f, g}
{f, c}
{e, c, f }
g
{b, e}
a
e
e
f
{a, b, e}
Figure 2: The graphs G(i) , G(j) and junction-trees T (i) and T (j) resulting from the removal
of the link Vij = {c, e}
of Vij = Vi ? Vj disconnects G into two (vertex-induced) subgraphs G(i) and G(j) which
are both triangulated, with junction-trees T (i) and T (j) respectively. We can recursively
decompose each of G(i) and G(j) into smaller and smaller subgraphs till the resulting subgraphs are cliques. When the size of all the minimal separators are bounded, we may use
these decompositions to easily solve problems that are hard in general. For example, in [5]
it is shown that NP-complete problems like vertex coloring, and finding maximum independent sets can be solved in polynomial time on graphs with bounded tree-width (which
are equivalent to spanning graphs with bounded size separators). We will be interested in
finding (triangulated) subgraphs of G that satisfy some conditions, such as a bound on the
number of edges, or a bound on the tree-width and which optimize separable objective
functions (described in Section 2)
One reason why problems such as this can often be solved easily when the tree-width of
G is bounded by some constant is this : If Vij is a separator decomposing G into G(i)
and G(j) , then a divide-and-conquer approach would suggest that we try and find optimal
subgraphs of G(i) and G(j) and then splice the two together to get an optimal subgraph of
G. There are two issues with this approach. First, the optimal subgraphs of G (i) and G(j)
need not necessarily match up on Vij , the set of common vertices. Second, even if the two
subgraphs agree on the set of common vertices, the graph resulting from splicing the two
subgraphs together need not be triangulated (which could happen even if the two subgraphs
individually are triangulated). To rectify the situation, we can do the following. We partition the set of subgraphs of G(i) and G(j) into classes, so that any subgraph of G(i) and any
subgraph G(j) corresponding to the same class are compatible in the sense that they match
up on their intersection namely Vij , and so that by splicing the two subgraphs together, we
get a subgraph of G which is acceptable (and in particular is triangulated). Then given optimal subgraphs of both G(i) and G(j) corresponding to each class, we can enumerate over
all the classes and pick the best one. Of course, to ensure that we do not repeatedly solve
the same problem, we need to work bottom-up (a.k.a dynamic programming) or memoize
our solutions. This procedure can be carried out in polynomial (in |V |) time as long as
we have only a polynomial number of classes. Now, if we have a polynomial number of
classes, these classes need not actually be a partition of all the acceptable subgraphs, though
the union of the classes must cover all acceptable subgraphs (so the same subgraph can be
contained in more than one class). For our application, every class can be thought of to be
the set of subgraphs that satisfy some constraint, and we need to pick a polynomial number
of constraints that cover all possibilities. The bound on the tree-width helps us here. If
k
|Vij | = k, then in any subgraph H of G, H[Vij ] must be one of the 2(2) possible subgraphs
k
of G[Vij ]. So, if k is sufficiently small (so 2(2) is bounded by some polynomial in |V |),
then this procedure results in a polynomial time algorithm. In this paper, we show that in
some cases we can characterize the space H so that we still have a polynomial number of
constraints even when the tree-width of G is not bounded by a small constant.
2.1 Separable objective functions
For cases where exact inference in the graphical model (G, PG ) is intractable, it is natural to
try to find a subgraphical model (H, PH ) such that D(PG kPH ) is minimized, and inference
using H is tractable. We will denote by H the set of subgraphs of G that are tractable for
inference. For example, this set could be the set of subgraphs of G with treewidth one less
than the treewidth of G, or perhaps the set of subgraphs of G with at d fewer edges. For a
specified subgraph H of G, there is a unique probability distribution PH factoring over H
that minimizes D(PG kPH ). Hence, finding a optimal subgraphical model is equivalent to
finding a subgraph H for which D(PG kPH ) is minimized. If Vij is a separator of G, we
will attempt to find optimal subgraphs of G by finding optimal subgraphs of G (i) and G(j)
and splicing them together. However, to do this, we need to ensure that the objective criteria
also decomposes along the separator Vij . Suppose that H is any triangulated subgraph of G.
Let PG(i) and PG(j) be the (marginalized) distributions of PG on V (i) and V (j) respectively,
and PH (i) and PH (j) be the (marginalized) distributions of the distribution PH on V (i) and
V (j) where H (i) = H[V (i) ] and H (j) = H[V (j) ], The following result assures us that the
KL-divergence also factors according to the separator Vij .
Lemma 1. Suppose that (G, PG ) is a graphical model, H is a triangulated subgraph of
G, and PH factors over H. Then D(PG kPH ) = D(PG(i) kPH (i) ) + D(PG(j) kPH (j) ) ?
D(PG[Vij ] kPH[Vij ] ).
Proof. Since H is a subgraph of G, and Vij is a separator of G, Vij must also be a sepa
P (i) ({Xv }v?V (i) )?PH (j) ({Xv }v?V (j) )
rator of H. Therefore, PH {Xv }v?V = H
. The result
)
PH[V ] ({Xv }
ij
v?Vij
follows immediately.
Therefore, there is hope that we can reduce our our original problem of finding an optimal
subgraph H ? H as one of finding subgraphs of H (i) ? G(i) and H (j) ? G(j) that are
compatible, in the sense that they match up on the overlap Vij , and for which D(PG kPH )
is minimized. Throughout this paper, for the sake of concreteness, we will assume that
the objective criterion is to minimize the KL-divergence. However, all the results can
be extended to other objective functions, as long as they ?separate? in the sense that for
any separator, the objective function is the sum of the objective functions of the two parts,
possibly modulo some correction factor which is purely a function of the separator. Another
example might be the complexity r(H) of representing the graphical model H. A very
natural representation satisfies r(G) = r(G(i) ) + r(G(j) ) if G has a separator G(i) ? G(j) .
Therefore, the representation cost reduction would satisfy r(G) ? r(H) = (r(G (i) ) ?
r(H (i) )) + (r(G(j) ) ? r(H (j) )), and so also factors according to the separators. Finally
note that any linear combinations of such separable functions is also separable, and so this
technique could also be used to determine tradeoffs (representation cost vs. KL-divergence
loss for example). In Section 4 we discuss some issues regarding computing this function.
2.2
Decompositions and decomposition trees
For the algorithms considered in this paper, we will be mostly interested in the decompositions that are specified by the junction tree, and we will represent these decompositions
by a rooted tree called a decomposition tree. This representation was introduced in [2, 3],
and is similar in spirit to Darwiche?s dtrees [6] which specify decompositions of directed
acyclic graphs. In this section and the next, we show how a decomposition tree for a graph
may be constructed, and show how it is used to solve a number of optimization problems.
abd; ce; gf
a; be; cd
d; bc; e
abe
dbc
ebc
e; cf ; g
cef
cf g
Figure 3: The separator tree corresponding to Figure 1
A decomposition tree for G is a rooted tree whose vertices correspond to separators and
cliques of G. We describe the construction of the decomposition tree in terms of a junctiontree T = (K, S) for G. The interior nodes of the decomposition tree R(T ) correspond to
S (the links of T and hence the minimal separators of G). The leaf or terminal nodes
represent the elements of K (the nodes of T and hence the maximal cliques of G). R(T )
can be recursively constructed from T as follows : If T consists of just one node K, (and
hence no edges), then R consists of just one node, which is given the label K as well. If
however, T has more than one node, then T must contain at least one link. To begin, let
Vi Vj ? S be any link in T . Then removal of the link Vi Vj results in two disjoint junctiontrees T (i) and T (j) . We label the root of R by the decomposition (V (i) ; Vij ; V (j) ). The rest
of R is recursively built by successively picking links of T (i) and T (j) (decompositions of
G(i) and G(j) ) to form the interior nodes of R. The effect of this procedure on the junction
tree of Figure 1 is shown in Figure 3, where the decomposition associated with the interior
nodes is shown inside the nodes. Let M be the set of all nodes of R(T ). For any interior
node M induced by the the link Vi Vj ? S of T , then we will let M (i) and M (j) represent
the left and right children of M , and R(i) and R(j) be the left and right trees below M .
3
3.1
Finding optimal subgraphical models
Optimal sub (k ? 1)-trees of k-trees
Suppose that G is a k-tree. A sub (k ? 1)-tree of G is a subgraph H of G that is (k ? 1)tree. Now, if Vij is any minimal separator of G, then both G(i) and G(j) are k-trees on
vertex sets V (i) and V (j) respectively. It is clear that the induced subgraphs H[V (i) ] and
H[V (j) ] are subgraphs of G(i) and G(j) and are partial (k ? 1)-trees. We will be interested
in finding sub (k ? 1)-trees of k trees and this problem is trivial by the result of [1] when
k = 2. Therefore, we assume that k ? 3. The following result characterizes the various
possibilities for H[Vij ] in this case.
Lemma 2. Suppose that G is a k-tree, and S = Vij is a minimal separator of G corresponding to the link ij of the junction-tree T . In any (k ? 1)-tree H ? G either
1. There is a u ? S such that u is not connected to vertices in both V (i) \ S and
V (j) \ S. Then S \ {u} is a minimal separator in H and hence is complete.
2. Every vertex in S is connected to vertices in both V (i) \S and V (j) \S. Then there
are vertices {x, y} ? S such that the edge H[S] is missing only the edge {x, y}.
Further either H[V (i) ] or H[V (j) ] does not contain a unchorded x-y path.
Proof. We consider two possibilities. In the first, there is some vertex u ? S such that u is
not connected to vertices in both V (i) \S and V (j) \. Since the removal of S disconnects G,
the removal of S must also disconnect H. Therefore, S must contain a minimal separator
of H. Since H is a (k ? 1)-tree, all minimal separators of H must contain k ? 1 vertices
which must therefore be S \{u}. This corresponds to case (1) above. Clearly this possiblity
can occur.
If there is no such u ? S, then every vertex in S is connected to vertices in both V (i) \ S
and V (j) \ S. If x ? S is connected to some yi ? V (i) \ S and yj ? V (j) \ S, then x is
contained in every minimal yi /yj separator (see [5]). Therefore, every vertex in S is part of
a minimal separator. Since each minimal separator contains k ? 1 vertices, there must be at
least two distinct minimum separators contained in S. Let Sx = S \ {x} and Sy = S \ {y}
be two distinct minimal separators. We claim that H[S] contains all edges except the edge
{x, y}. To see this, note that if z, w ? S, with z 6= w and {z, w} 6= {x, y} (as sets),
then either {z, w} ? Sy or {z, w} ? Sx . Since both Sx and Sy are complete in H, this
edge must be present in H. The edge {x, y} is not present in H[S] because all minimal
separators in H must be of size k ? 1. Further, if both V (i) and V (j) contain an unchorded
path between x and y, then by joining the two paths at x and y, we get a unchorded cycle
in H which contradicts the fact that H is triangulated.
Therefore, we may associate k2 ? 2 + 2 ? k constraints with each separator Vij of G as
follows. There are k possible constraints corresponding to case (1) above (one for each
choice of x), and k2 ? 2 choices corresponding to case (2) above. This is because for each
pair {x, y} corresponding to the missing edge, we have either V (i) contains no unchorded
xy paths or V (j) contains no unchorded xy paths. More explicitly, we can encode the set
of constraints CM associated with each separator S corresponding to an interior node M of
the decomposition tree as follows: CM = { (x, y, s) : x ? S, y ? S, s ? {i, j}}. If y = x,
then this corresponds to case (1) of the above lemma. If s = i, then x is connected only to
H (i) and if s = j, then x is connected only to H (j) . If y 6= x, then this corresponds to case
(2) in the above lemma. If s = i, then H (i) does not contain any unchorded path between
x and y, and there is no constraint on H (j) . Similarly if s = j, then H (j) does not contain
any unchorded path between x and y, and there is no constraint on H (i) .
Now suppose that H (i) and H (j) are triangulated subgraphs of G(i) and G(j) respectively,
then it is clear that if H (i) and H (j) both satisfy the same constraint they must match up
on the common vertices Vij . Therefore to splice together two solutions corresponding to
the same constraint, we only need to check that the graph obtained by splicing the graphs
is triangulated.
Lemma 3. Suppose that H (i) and H (j) are triangulated subgraphs of G(i) and G(j) respectively such that both of them satisfy the same constraint as described above. Then the
graph H obtained by splicing H (i) and H (j) together is triangulated.
Proof. Suppose that both H (i) and H (j) are both triangulated and both satisfy the same
constraint. If both H (i) and H (j) satisfy the same constraint corresponding to case (1)
in Lemma 2 and H has an unchorded cycle, then this cycle must involve elements of both
H (i) and H (j) . Therefore, there must be two vertices of S \{u} on the cycle, and hence this
cycle has a chord as S \ {u} is complete. This contradiction shows that H is triangulated.
So assume that both of them satisfy the constraint corresponding to case (2) of Lemma 2.
Then if H is not triangulated, there must be a t-cycle (for t ? 4) with no chord. Now, since
{x, y} is the only missing edge of S in H, and because H (i) and H (j) are individually
triangulated, the cycle must contain x, y and vertices of both V (i) \ S and V (j) \ S. We
may split this unchorded cycle into two unchorded paths, one contained in V (i) and one in
V (j) thus violating our assumption that both H (i) and H (j) satisfy the same constraint.
If |S| = k, then there are 2k + 2 ? k2 ? O(k 2 ) ? O(n2 ). We can use a divide and conquer
strategy to find the optimal sub (k ? 1) tree once we have taken care of the base case, where
G is just a single clique (of k + 1) elements. However, for this case, it is easily checked that
any subgraph of G obtained by deleting exactly one edge results in a (k ? 1) tree, and every
sub (k?1)-tree results from this operation. Therefore, the optimal (k?1)-tree can be found
using Algorithm 1, and in this case, the complexity of Algorithm 1 is O(n(k + 1) 2 ). This
procedure can be generalized to find the optimal sub (k ? d)- tree for any fixed d. However,
the number of constraints grows exponentially with d (though it is still polynomial in n).
Therefore for small, fixed values of d, we can compute the optimal sub (k ? d)-tree of G.
While we can compute (k ? d)-trees of G by first going from a k tree to a (k ? 1) tree, then
from a (k ? 1)-tree to a (k ? 2)-tree, and so on in a greedy fashion, this will not be optimal
in general. However, this might be a good algorithm to try when d is large.
3.2
Optimal triangulated subgraphs with |E(G)| ? d edges
Suppose that we are interested in a (triangulated) subgraph of G that contains d fewer edges
that G does. That is, we want to find an optimal subgraph H ? G such that |E(H)| =
|E(G)| ? d. Note that by the result of [4] there is always a triangulated subgraph with d
fewer edges (if d < |E(G)|). Two possibilities for finding such an optimal subgraph are
1. Use the procedure described in [4]. This is a greedy procedure which works in
d steps by deleting an edge at each step. At each state, the edge is picked from
the set of edges whose deletion leaves a triangulated graph. Then the edge which
causes the least increase in KL-divergence is picked at each stage.
2. For each possible subset A of E(G) of size d, whose deletion leaves a triangulated
graph, compute the KL divergence using
the formula above, and then pick the
|E(G)|
such sets, this can be done in polynomial
optimal one. Since there are
d
time (in |V (G)|) when d is a constant.
The first greedy algorithm is not guaranteed to yield the optimal solution. The second takes
time that is O(n2d ). Now, let us solve this problem using the framework we?ve described.
Let H be the set of subgraphs of G which may be obtained by deletion of d
edges.
For each M = ij ? M corresponding to the o
separator Vij , let CM =
n
E(G[Vij ])
(l, r, c, s, A) : l + r ? c = d, s a d bit string, A ?
. The constraint reprec
sented by (l, r, c, A) is this : A is a set of d edges of G[Vij ] that are missing in H, l edges
are missing from the left subgraph, and r edges are missing from the right subgraph. c represents the double count, and so is subtracted from the total. If k is the size of the largest
k
clique, then the total number of such constraints is bounded by 2d ? 2d ? (2) ? O(k 2d )
d
which could be better than O(n2d ) and is polynomial in |V | when d is constant. See [10]
for additional details.
4
Conclusions
Algorithm 1 will compute the optimal H ? H for the two examples discussed above and
is polynomial (for fixed constant d) even if k is O(n). In [10] a generalization is presented
which will allow finding the optimal solution for other classes of subgraphical models.
Now, we assume an oracle model for computing KL-divergences of probability distributions on vertex sets of cliques. It is clear that these KL-divergences can be computed
R ? separator-tree for G;
for each vertex M of R in order of increasing height (bottom up) do
for each constraint cM of M do
if M is an interior vertex of R corresponding to edge ij of the junction tree then
Let Ml and Mr be the left and right children of M ;
Pick constraint cl ? CMl compatible with cM to minimize table[Ml , cl ];
Pick constraint cr ? CMr compatible with cM to minimize table[Mr , cr ];
loss ? D(PG [M ]kPH [M ]);
table[M, cM ] ? table[Ml , cl ] + table[Mr , cr ] ? loss;
else
table[M, cM ] ? D(PG [M ]kPH [M ]);
end
end
end
Algorithm 1: Finding optimal set of constraints
efficiently for distributions like Gaussians, but for discrete distributions this may not be
possible when k is large. However even in this case this algorithm will result in only
polynomial calls to the oracle. The standard algorithm [3] which is exponential in the
treewidth will make O(2k ) calls to this oracle. Therefore, when the cost of computing the
KL-divergence is large, this algorithm becomes even more attractive as it results in exponential speedup over the standard algorithm. Alternatively, if we can compute approximate
KL-divergences, or approximately optimal solutions, then we can compute an approximate
solution by using the same algorithm.
References
[1] C. Chow and C. Liu, ?Approximating discrete probability distributions with dependence trees?, IEEE Transactions on Information Theory, v. 14, 1968, Pages 462?467.
[2] F. Gavril, ?Algorithms on clique separable graphs?, Discrete Mathematics v. 9 (1977),
pp. 159?165.
[3] R. E. Tarjan. ?Decomposition by Clique Separators?, Discrete Mathematics, v. 55
(1985), pp. 221?232.
[4] U. Kjaerulff. ?Reduction of computational complexity in Bayesian networks through
removal of weak dependencies?, Proceedings of the Tenth Annual Conference on
Uncertainty in Artificial Intelligence, pp. 374?382, 1994.
[5] T. Kloks, ?Treewidth: Computations and Approximations?, Springer-Verlag, 1994.
[6] A. Darwiche and M. Hopkins. ?Using recursive decomposition to construct elimination orders, jointrees and dtrees?, Technical Report D-122, Computer Science Dept.,
UCLA.
[7] S. Lauritzen. ?Graphical Models?, Oxford University Press, Oxford, 1996.
[8] T. A. McKee and F. R. McMorris. ?Topics in Intersection Graph Theory?, SIAM
Monographs on Discrete Mathematics and Applications, 1999.
[9] D. Karger and N. Srebro. ?Learning Markov networks: Maximum bounded tree-width
graphs.? In Symposium on Discrete Algorithms, 2001, Pages 391-401.
[10] M. Narasimhan and J. Bilmes. ?Optimization on separator-clique trees.?, Technical
report UWEETR 2004-10, June 2004.
| 2682 |@word polynomial:17 dtrees:2 memoize:1 seek:1 sepa:1 decomposition:19 cml:1 pg:18 pick:5 recursively:3 reduction:2 liu:1 contains:6 selecting:1 karger:1 bc:1 existing:1 must:16 happen:1 partition:2 v:1 greedy:4 fewer:4 leaf:3 intelligence:1 node:15 height:1 along:1 constructed:2 symposium:1 consists:2 inside:1 darwiche:2 terminal:1 increasing:1 becomes:1 begin:1 bounded:10 notation:1 cm:8 minimizes:1 string:1 narasimhan:2 finding:15 corporation:1 impractical:1 guarantee:1 every:7 growth:1 exactly:1 k2:3 grant:2 engineering:1 xv:4 joining:1 oxford:2 path:8 approximately:1 might:2 directed:1 unique:1 yj:2 union:1 recursive:1 procedure:6 rator:1 thought:1 suggest:1 get:3 cannot:2 interior:6 optimize:1 equivalent:2 missing:6 immediately:1 subgraphs:30 contradiction:1 construction:1 suppose:8 modulo:1 exact:3 programming:2 us:1 associate:1 element:5 observed:1 bottom:2 electrical:1 solved:3 ensures:1 cycle:9 connected:7 chord:3 monograph:1 complexity:8 dynamic:2 abd:1 purely:1 easily:3 joint:1 various:1 distinct:3 describe:1 artificial:1 quite:1 whose:3 solve:5 say:1 maximal:3 remainder:1 subgraph:23 till:1 seattle:1 empty:2 double:1 help:1 ij:4 lauritzen:1 treewidth:8 triangulated:27 closely:1 elimination:1 generalization:1 preliminary:1 decompose:1 annihilation:1 extension:1 correction:1 sufficiently:1 considered:1 claim:1 reserve:1 combinatorial:1 label:3 individually:2 largest:3 hope:1 clearly:1 always:1 cr:3 encode:1 june:1 properly:1 check:1 sense:3 inference:9 factoring:1 typically:1 chow:1 going:1 interested:4 issue:2 special:1 once:1 construct:1 washington:2 represents:1 minimized:3 np:3 report:2 divergence:13 ve:1 attempt:1 investigate:1 possibility:4 subtrees:1 n2d:2 edge:28 partial:1 xy:2 tree:60 divide:3 minimal:19 cover:2 cost:3 introducing:1 vertex:28 subset:2 characterize:1 dependency:1 siam:1 picking:1 together:6 quickly:1 hopkins:1 successively:1 possibly:1 disconnect:4 satisfy:9 explicitly:1 vi:13 try:3 root:1 picked:2 characterizes:1 kph:10 minimize:4 efficiently:1 sy:3 correspond:5 yield:1 weak:1 bayesian:1 bilmes:3 checked:1 pp:3 proof:3 associated:2 actually:1 coloring:1 violating:1 specify:1 formulation:1 done:1 though:2 just:3 stage:1 perhaps:1 grows:2 effect:1 contain:8 dbc:1 hence:8 attractive:1 adjacent:2 width:8 rooted:2 criterion:2 generalized:1 complete:7 common:3 mckee:1 exponentially:1 discussed:1 jointrees:1 mathematics:3 similarly:1 rectify:1 base:1 ebc:1 verlag:1 arbitrarily:1 yi:2 minimum:1 greater:1 care:1 additional:1 mr:3 determine:1 ii:1 technical:2 match:4 long:2 impact:1 represent:3 want:1 else:1 rest:1 induced:4 spirit:1 call:2 ee:1 split:1 rendering:1 restrict:1 reduce:1 regarding:2 tradeoff:1 possiblity:1 cause:1 repeatedly:1 enumerate:1 clear:3 involve:1 ph:13 nsf:1 disjoint:1 discrete:7 independency:1 terminology:1 ce:1 tenth:1 graph:25 concreteness:1 sum:1 uncertainty:1 throughout:1 splicing:5 separation:1 sented:1 acceptable:3 bit:1 bound:3 guaranteed:1 oracle:3 annual:1 occur:1 constraint:22 x2:2 sake:1 ucla:1 optimality:1 separable:5 speedup:1 according:3 combination:1 smaller:3 contradicts:1 taken:1 agree:1 assures:1 discus:1 count:1 tractable:3 end:3 junction:12 decomposing:1 operation:1 gaussians:1 subtracted:1 original:2 ensure:2 cf:2 graphical:15 marginalized:2 conquer:3 approximating:1 objective:7 strategy:1 dependence:1 said:2 link:12 separate:1 topic:1 trivial:1 spanning:2 reason:1 assuming:1 cmr:1 length:1 gavril:2 mostly:1 proper:1 markov:1 defining:1 situation:1 extended:1 arbitrary:1 tarjan:2 abe:1 introduced:1 pair:2 required:1 kl:13 specified:3 namely:1 deletion:3 below:1 built:1 including:1 deleting:2 overlap:1 natural:2 solvable:1 representing:1 imply:1 carried:1 gf:1 removal:8 loss:3 expect:1 acyclic:1 srebro:1 kjaerulff:1 vij:28 cd:1 compatible:4 course:1 supported:1 allow:1 xn:2 simplified:1 transaction:1 approximate:2 clique:17 ml:3 alternatively:1 decomposes:1 why:1 table:6 nature:1 necessarily:1 separator:44 cl:3 vj:12 n2:1 child:2 x1:2 intel:1 fashion:1 sub:8 exponential:4 splice:2 formula:1 mukund:1 intractable:1 subtree:1 sx:3 intersection:2 contained:5 applies:2 springer:1 corresponds:3 satisfies:1 conditional:1 jeff:1 hard:1 except:1 reducing:2 lemma:7 called:2 total:2 dept:2 |
1,853 | 2,683 | Rate- and Phase-coded Autoassociative Memory
M?t? Lengyel
Peter Dayan
Gatsby Computational Neuroscience Unit, University College London
17 Queen Square, London WC1N 3AR, United Kingdom
{lmate,dayan}@gatsby.ucl.ac.uk
Abstract
Areas of the brain involved in various forms of memory exhibit patterns
of neural activity quite unlike those in canonical computational models.
We show how to use well-founded Bayesian probabilistic autoassociative
recall to derive biologically reasonable neuronal dynamics in recurrently
coupled models, together with appropriate values for parameters such as
the membrane time constant and inhibition. We explicitly treat two cases.
One arises from a standard Hebbian learning rule, and involves activity
patterns that are coded by graded firing rates. The other arises from a
spike timing dependent learning rule, and involves patterns coded by the
phase of spike times relative to a coherent local field potential oscillation.
Our model offers a new and more complete understanding of how neural
dynamics may support autoassociation.
1
Introduction
Autoassociative memory in recurrently coupled networks seems fondly regarded as having been long since solved, at least from a computational perspective. Its neurobiological
importance, as a model of episodic (event) memory storage and retrieval (from noisy and
partial inputs) in structures such as area CA3 in the hippocampus, is of course clear [1].
This perhaps suggests that it is only the exact mapping of the models to the neural substrate
that holds any remaining theoretical interest.
However, the characteristic patterns of activity in areas such as CA3 that are involved in
memory are quite unlike those specified in the bulk of models. In particular neurons (for
instance hippocampal place cells) show graded activity during recall [2], prominent theta
frequency oscillations [3] and an apparent variety of rules governing synaptic plasticity
[4, 5]. The wealth of studies of memory capacity of attractor networks of binary units does
not give many clues to the specification, analysis or optimization of networks acting in
these biologically relevant regimes. In fact, even theoretical approaches to autoassociative
memories with graded activities are computationally brittle.
Here, we generalize previous analyses [6, 7] to address these issues. Formally, these models interpret recall as Bayesian inference based on information given by the noisy input,
the synaptic weight matrix, and prior knowledge about the distribution of possible activity patterns coding for memories. More concretely (see section 2), the assumed activity
patterns and synaptic plasticity rules determine the term in neuronal update dynamics that
describes interactions between interconnected cells. Different aspects of biologically reasonable autoassociative memories arise from different assumptions. We show (section 3)
We thank Boris Gutkin for helpful discussions on the phase resetting characteristics of different
neuron types. This work was supported by the Gatsby Charitable Foundation.
that for neurons are characterized by their graded firing rates, the regular rate-based characterization of neurons effectively approximates optimal Bayesian inference. Optimal values
for parameters of the update dynamics, such as level of inhibition or leakage conductance,
are inherently provided by our formalism. We then extend the model (section 4) to a setting involving spiking neurons in the context of a coherent local field potential oscillation
(LFPO). Memories are coded by the the phase of the LFPO at which each neuron fires, and
are stored by spike timing dependent plasticity. In this case, the biophysically plausible
neuronal interaction function takes the form of a phase reset curve: presynaptic firing accelerates or decelerates the postsynaptic cell, depending on the relative timing of the two
spikes, to a degree that is proportional to the synaptic weight between the two cells.
2
MAP autoassociative recall
The first requirement is to specify the task for autoassociative recall in a probabilistically
sound manner. This specification leads to a natural account of the dynamics of the neurons
during recall, whose form is largely determined by the learning rule. Unfortunately, the
full dynamics includes terms that are not purely local to the information a post-synaptic
neuron has about pre-synaptic activity, and we therefore consider approximations that restore essential characteristics necessary to satisfy the most basic biological constraints. We
validate the quality of the approximations in later sections.
The construction of the objective function: Consider an autoassociative network which
has stored information about M memories x1 . . . xM in a synaptic weight matrix, W between a set of N neurons. We specify these quantities rather generally at first to allow for
different ways of construing the memories later. The most complete probabilistic description of its task is to report the conditional distribution P [x|?
x, W] over the activities x given
noisy inputs x
? and the weights. The uncertainty in this posterior distribution has two roots.
First, the activity pattern referred to by the input is unclear unless there is no input noise.
Second, biological synaptic plasticity rules are data-lossy ?compression algorithms?, and
so W specifies only imprecise information about the stored memories.
In an ideal case, P [x|?
x, W] would have support only on the M stored patterns x1 . . . xM .
However, biological storage methods lead to weights W that permit a much greater range
of possibilities. We therefore consider methods that work in the full space of activities x.
In order to optimize the probability of extracting just the correct memory, decision theory
encourages us to maximize the posterior probability [8]:
x
? := argmaxx P [x|?
x, W] , P [x|?
x, W] ? P [x] P [?
x|x] P [W|x]
(1)
The first term in Eq.1 imports prior knowledge
Q of the statistical characteristics of the memories, and is assumed to factorize: P [x] := i Px [xi ]. The second term describes the noise
process corrupting the inputs. For unbiased noise it will be a term in x that is effectively
centered on x
?. We assume that the noise corrupting each element
Q of the patterns
Q is independent, and independent of the original pattern, so P [?
x|x] := i P [?
xi |x] := i P [?
xi |xi ].
The third term assesses the likelihood that the weight matrix came from a training set of
size M including pattern x.1 Biological constraints encourage consideration of learning
updates for the synapse from neuron j to neuron i that are local to the pre-synaptic (xm
j )
m
and post-synaptic (xm
i ) activities of connected neurons when pattern x is stored:
m
m
?wi,j
:= ? xm
(2)
i , xj
We assume the contributions of individual training patterns are additive, Wi,j :=
P
m
m ?wi,j , and that there are no autapses in the network, Wi,i := 0.
1
Uncertainty about M could also be incorporated into the model, but is neglected here.
Storing a single random pattern drawn from the prior distribution will result in a synaptic
weight change with a distribution determined by the prior
and the learning rule, having
2
??w = h? (x1 , x2 )iPx [x1 ]?Px [x2 ] mean, and ??w
= ?2 (x1 , x2 ) P [x ]?P [x ] ? ?2?w
x 1
x 2
variance. Storing M ? 1 random patterns means adding M ? 1 iid. random variables and
thus, for moderately large M , results in a synaptic weight with an approximately Gaussian
distribution P [Wi,j ] ' G (Wi,j ; ?W , ?W ), with mean ?W = (M ? 1) ??w and variance
2
2
?W
= (M ? 1) ??w
. Adding a further particular pattern x is equivalent to adding a
random variable with a mean determined by the learning rule, and zero variance, thus:
P [Wi,j |xi , xj ] ' G (Wi,j ; ?W + ? (xi , xj ) , ?W )
(3)
We also make the approximation that
Q elements of the synaptic weight matrix are independent, and thus write: P [W|x] := i,j6=i P [Wi,j |xi , xj ].
Having restricted our horizons to maximum a posteriori (MAP) inference, we can consider
as an objective function the log of the posterior distribution. In the light of our factorizability assumptions, this is
O (x) = log
[?
x|x] + log P [W|x]
P P [x] + log PP
P
= i log P [xi ] + i log P [?
xi |xi ] + i,j6=i log P [Wi,j |xi , xj ]
(4)
Neuronal update dynamics: Finding the global maximum of the objective function, as
stated in equation 1, is computationally extravagant, and biologically questionable. We
therefore specify neuronal dynamics arising from gradient ascent on the objective function:
?x x? ? ?x O (x) .
(5)
Combining equations 4 and 5 we get
i
?x dx
dt =
?
?xi
log P [W|x] =
?
?xi
log P [x] +
?
j6=i ?xi
P
?
?xi
?
?xi log P [W|x] , where
?
?xi log P [Wj,i |xj , xi ] .
log P [?
x|x] +
log P [Wi,j |xi , xj ] +
(6)
(7)
The first two terms in equation 6 only depend on the activity of the neuron itself and its
input. For example, for a Gaussian prior Px [xi ] = G (Wi,j ; ?x , ?x ) and unbiased Gaussian
noise on the input P [?
xi |xi ] = G (?
xi ; xi , ?x? ), these would be:
d
d
i
i
= ??x2 ? ?12 + ?12 xi + ?x?2i (8)
xi |xi ] = ?x??x
+ x?i??x
2
2
dxi log P [xi ] + dxi log P [?
x
x
?
x
x
x
?
x
?
The first term on the right-hand side of the last equality expresses a constant bias; the
second involves self-decay; and the third describes the effect of the input.
The terms in equation 7 indicate how a neuron should take into account the activity of other
neurons based on the synaptic weights. From equation 3, the terms are
?
?xi
?
?xi
log P [Wi,j |xi , xj ]=
log P [Wj,i |xj , xi ]=
1
2
?W
h
(Wi,j ? ?W )
i
(9)
1
2
?W
h
i
?
?
(Wj,i ? ?W ) ?x
?(xj , xi ) ? ?(xj , xi ) ?x
?(xj , xi )
i
i
(10)
?
?(xi , xj )
?xi
? ?(xi , xj )
?
?(xi , xj )
?xi
Two aspects of the above formul? are biologically troubling. The last terms in each express the effects of other cells, but without there being corresponding synaptic weights. We
approximate these terms using their mean values over the prior distribution. In this case
?
?
?i+ = h? (xi , xj ) ?x
? (xi , xj )iPx [xj ] and ?i? = h? (xj , xi ) ?x
? (xj , xi )iPx [xj ] coni
i
tribute terms that only depend on the activity of the updated cell, and so can be lumped
with the prior- and input-dependent terms of Eq.8.
Further, equation 10 includes synaptic weights, Wj,i , that are postsynaptic with respect
to the updated neuron. This would require the neuron to change its activity depending
on the weights of its postsynaptic synapses. One simple work-around is to approximate
a postsynaptic weight by the mean of its conditional distribution given the corresponding
presynaptic weight: Wj,i ' hP [Wj,i |Wi,j ]i. In the simplest case of perfectly symmetric
or anti-symmetric learning, with ? (xi , xj ) = ?? (xj , xi ), we have Wj,i = ?Wj,i and
?i+ = ?i? = ?i . In the anti-symmetric case ?w = 0.
Making these assumptions, the neuronal interaction function simplifies to
?
? (xi , xj )
(11)
H (xi , xj ) = (Wi,j ? ?W ) ?x
i
i
and ?22
j6=i H (xi , xj ) ? (N ? 1) ?i is the weight-dependent term of equation 7.
W
Equation 11 shows that there is a simple relationship between the synaptic plasticity rule,
? (xi , xj ), and the neuronal interaction function, H (xi , xj ), that is approximately optimal
for reading out the information that is encoded in the synaptic weight matrix by that synaptic plasticity rule. It also shows that the magnitude of this interaction should be proportional
to the synaptic weight connecting the two cells, Wi,j .
hP
We specialize this analysis to two important cases with (a) graded, rate-based, or (b) spiking, oscillatory phase-based, activities. We derive appropriate dynamics from learning
rules, and show that, despite the approximations, the networks have good recall performance.
3
Rate-based memories
The most natural assumption about pattern encoding is that the activity of each unit is
interpreted directly as its firing rate. Note, however, that most approaches to autoassociative
memory assume binary patterns [9], sitting ill with the lack of saturation in cortical or
hippocampal neurons in the appropriate regime. Experiments [10] suggest that regulating
activity levels in such networks is very tricky, requiring exquisitely carefully tuned neuronal
dynamics. There has been work on graded activities in the special case of line or surface
attractor networks [11, 12], but these also pose dynamical complexitiese. By contrast,
graded activities are straightforward in our framework.
Consider Hebbian covariance learning: ?cov (xi , xj ) := Acov (xi ? ?x ) (xj ? ?x ), where
Acov > 0 is a normalizing constant and ?x is the mean of the prior distribution of the
patterns to be stored. The learning rule is symmetric, and so, based on Eq.11, the optimal
neuronal interaction function is Hcov (xi , xj ) = Acov (Wi,j ? ?W ) (xj ? ?x ). This leads
to a term in the dynamics which is the conventional weighted sum of pre-synaptic firing
rates. The other key term in the dynamics is ?i = ?A2cov ?x2 (xi ? ?x ), where ?x2 is the
variance of the prior distribution, expressing self-decay to a baseline activity level determined by ?x . The prior- and input-dependent terms also contribute to self-decay as shown
in Eq.8. Integration of the weighted sum of inputs plus decay to baseline constitute the
widely used leaky integrator reduction of a single neuron [10].
Thus, canonical models of synaptic plasticity (the Hebbian covariance rule) and single neuron firing rate dynamics are exactly matched for autoassociative recall. Optimal values for
all parameters of single neuron dynamics (except the membrane time constant determining the speed of gradient ascent) are directly implied. This is important, since it indicates
how to solve the problem for graded autoassociative memories (as opposed to saturing ones
[14, 15]), that neuronal dynamics have to be finely tuned. As examples, the leak conductance is given by the sum of the coefficients of all terms linear in xi , the optimal bias current
is the sum of all terms independent of xi , and the level of inhibition can be determined from
the negative terms in the interaction function, ??W and ??x .
Since our derivation embodies a number of approximations, we performed numerical simulations. To gauge the performance of the Bayes-optimal network we compared it to
networks of increasing complexity (Fig. 1A,B). A trivial lower bound of performance is
A
B
prior
input
ideal observer
Bayesian: prior + input
Bayesian: prior + input + synapses
3
2
0.7
prior
input
Bayesian: prior + input
Bayesian: prior + input + synapses
1
0.5
Frequency
Recalled activity
0.6
0
?1
0.4
0.3
0.2
?2
0.1
?3
?3
?2
?1
0
1
Stored activity
2
?4
?3
?2
?1
0
Error
1
2
3
4
5
D
C
1.2
Average normalized error
1.2
prior
input
ideal observer
Bayesian
Treves
1.4
prior
input
ideal observer
Bayesian
Treves
1.4
1
Average error
0
?5
3
0.8
0.6
0.4
1
0.8
0.6
0.4
0.2
0.2
0
0
1
10
Number of stored patterns
100
1
10
Number of stored patterns
100
Figure 1: Performance of the rate-coded Bayesian inference network (), compared to a Bayesian
network that only takes into account evidence from the prior and the input but not from the synaptic
weight matrix (?), a network that randomly generates patterns from the prior distribution (?), a
network that transmits its input to its output (+), and the ?ideal observer? having access to the list
of stored patterns (?). A. Firing rates of single units at the end of the recall process (y-axis) against
firing rates in the original pattern (x-axis). B. Frequency histograms of errors (difference between
recalled and stored firing rates). The ideal observer is not plotted because its error distribution was a
Dirac-delta at 0. C, D. Benchmarking the Bayesian network against the network of Treves [13] (?)
on patterns of non-negative firing rates. Average error is the square root of the mean squared error
(C), average normalized error measures only the angle difference between true and recalled activities
(D). (These measures are not exactly the same as that used to derive the dynamics (equation 1), but
are reasonably appropriate.) The prior distribution was Gaussian with ?x = 0 mean and ?x2 = 1
variance (A,B), or a Gaussian with ?x = 0.5 mean and ?x2 = 0.25 variance truncated below 0 (C)
(yielding approximately a = 0.5 density), or ternary with a = 0.5 mean and density (D). The input
was corrupted by unbiased Gaussian noise of ?x2? = 1 variance (A,B), or ?x2? = 1.5 variance (C,D) and
cut at 0 (C,D). The learning rule was the covariance rule with Acov = 1 (A,B), or with Acov = 1/N a2
(C,D). The number of cells in the network was N = 50 (A,B) and N = 100 (C,D), and the number of
memories stored was M = 2 (A,B) or varied between M = 2 . . . 100 (C,D, note logarithmic scale).
For each data point, 10 different networks were simulated with a different set of stored patterns, and
for each network, 10 attempts at recall were made, with a noisy version of a randomly chosen pattern
as the input and with activities initialized at this input.
given by a network that generates random patterns from the same prior distribution from
which the patterns to be stored were drawn (P [x]). Another simple alternative is a network that simply transmits its input (?
x) to its output. (Note that the ?input only? network
is not necessarily superior to the ?prior only? network: their relative effectiveness depends
on the relative variances of the prior and noise distributions, a narrow prior with a wide
noise distribution would make the latter perform better, as in Fig. 1D). The Bayesian inference network performs considerably better than any of these simple networks. Crucially,
this improvement depends on the information encoded in synaptic weights: the network
practically falls back to the level of the ?input only? network (or the ?prior only? network,
whichever is the better, data not shown) if this information is ignored at the construction of
the recall dynamics (by taking the third term in Eq. 6 to be 0).
An upper bound on the performance of any network using some biological form of synaptic
plasticity comes from an ?ideal observer? which knows the complete list of stored patterns
(rather than its distant reflection in the synaptic weight matrix) and computes and compares
the probability that each was corrupted to form the input x
? to find the best match (rather
than using neural dynamics). Such an ideal observer only makes errors when both the
number of patterns stored and the noise in the input is sufficiently large, so that corrupting
a stored pattern is likely to make it more similar to another stored pattern. In the case
shown in Fig. 1A,B, this is not the case, since only two patterns were stored, and the ideal
observer performs perfectly as expected. Nevertheless, there may be situations in which
perfect performance is out of reach even for an ideal observer (Fig. 1C,D), which makes it
a meaningful touchstone. In summary, the performance of any network can be assessed by
measuring where it lies between the better one of the ?prior only? and ?input only? networks
and the ideal observer.
As a further challenge, we also benchmarked our model against the model of Treves [13]
(Fig. 1C,D), which we chose because it is a rare example of a network that was designed to have near optimal recall performance in the face of non-binary patterns. In
this work,
drawn from the distribution P [xi ] :=
Treves considered ternary patterns,
1 ? 43 a ? (xi ) + a? xi ? 21 + a3 ? xi ? 32 , where ? (x) is the Dirac-delta function.
Here, a = ?x quantifies the density of the patterns (i.e. how non-sparse they are). The
patterns are stored using the covariance rule as stated above (with Acov := N1a2 ). Neuronal
update in the model is discrete, asynchronous, and involves two steps. First the ?local field?
P
P
3
is calculated as hi := j6=i Wi,j xj ? k ( i xi ? N ) + Input, then the output of the
neuron is calculated as a threshold linear function of the local field: xi := g (hi ? hThr )
if hi > hThr and xi := 0 otherwise, where g := 0.53 a/ (1 ? a) is the gain parameter,
and hThr := 0 is the threshold, and the value of k is set by iterative search to optimize
performance.
The comparison between Treves? network as we implemented it and our network is imperfect, since the former is optimized for recalling ternary patterns while, in the absence of
neural evidence for ternary patterns, we used the simpler and more reasonable neural dynamics for our network that emerge from an assumption that the distribution over the stored
patterns is Gaussian. Further, we corrupted the inputs by unbiased additive Gaussian noise
(with variance ?x2? = 1.5), but truncated the activities at 0, though did not adjust the dynamics of our network in the light of the truncation. Of course, these can only render our
network less effective. Still, the Bayesian network clearly outperformed the Treves network
when the patterns were drawn from a truncated Gaussian (Fig. 1C). The performance of the
Bayesian network stayed close to that of an ideal observer assuming non-truncated Gaussian input, showing that most of the errors were caused by this assumption and not from
suboptimality of neural interactions decoding the information in synaptic weights. Despite
extensive efforts to find the optimal parameters for the Treves network, its performance did
not even reach that of the ?input only? network.
Finally, again for ternary patterns, we also considered only penalizing errors about the direction of the vectors of recalled activities ignoring errors about their magnitudes (Fig. 1D).
The Treves network did better in this case, but still not as well as the Bayesian network.
Importantly, in both cases, in the regime where synaptic weights were saturated in the
M ? N limit and thus it was no longer possible to extract any useful information from the
synaptic weights, the Bayesian network still only fell back to the level of the ?prior only?
network, but the Treves network did not seem to have any such upper bound on its errors.
4
Phase-based memories
Brain areas known to be involved in memory processing demonstrate prominent oscillations (LFPOs) under a variety of conditions, including both wake and sleep states [16].
Under these conditions, the phases of the spikes of a neuron relative to the LFPO have been
shown to be carefully controlled [17], and even to convey meaningful stimulus information,
e.g. about the position of an animal in its environment [3] or retrieved odor identity [18].
The discovery of spike timing dependent plasticity (STDP) in which the relative timing
of pre- and postsynaptic firings determines the sign and extent of synaptic weight change,
offered new insights into how the information represented by spike times may be stored
in neural networks [19]. However, bar some interesting suggestions about neuronal resonance [20], it is less clear how one might correctly recall information thereby stored in the
synaptic weights.
The theory laid out in Section 2 allows us to treat this problem systematically. First,
neuronal activities, xi , will be interpreted as firing times relative to a reference phase
of the ongoing LFPO, such as the peak of theta oscillation in the hippocampus, and
will thus be circular variables drawn from a circular Gaussian. Next, our learning rule
is an exponentially decaying Gabor-function of the phase difference between pre- and
postsynaptic firing: ?STDP (xi , xj ) := ASTDP exp[?STDP cos(??i,j )] sin(??i,j ? ?STDP )
with ??i,j = 2? (xi ? xj ) /TSTDP . STDP characteristics in different brain regions are
well captured by this general formula, but the parameters determining their exact shapes
greatly differ among regions. We constrain our analysis to the antisymmetric case, so
that ?STDP = 0, and set other parameters to match experimental data on hippocampal
STDP [5]. The neuronal interaction function
that satisfies Eq.11 is HSTDP (x
i , xj ) =
2?ASTDP /TSTDP Wi,j exp[?STDP cos(??i,j )] cos(??i,j ) ? ?STDP sin2 (??i,j ) . This interaction function decreases firing phase, and thus accelerates the postsynaptic cell if the
presynaptic spike precedes postsynaptic firing, and delays the postsynaptic cell if the presynaptic spike arrives just after the postsynaptic cell fired. This characteristic is the essence
of the biphasic phase reset curve of type II cells [21], and has been observed in various
types of neurons, including neocortical cells [22]. Thus again, our derivation directly couples STDP, a canonical model of synaptic plasticity, and phase reset curves in a canonical
model of neural dynamics.
Numerical simulations tested again the various approximations. Performance of the network is shown in Fig.2 as is comparable to that of the rate coded network (Fig.2). Further
simulations will be necessary to map out the performance of our network over a wider
range of parameters, such as the signal-to-noise ratio.
5
Discussion
We have described a Bayesian approach to recall in autoassociative memories. This permits
the derivation of neuronal dynamics appropriate to a synaptic plasticity rule, and we used
this to show a coupling between canonical Hebbian and STDP plasticity rules and canonical
rate-based and phase-based neuronal dynamics respectively. This provides an unexpectedly
close link between optimal computations and actual implementations. Our method also
leads to networks that are highly competent at recall.
There are a number of important direction for future work. First, even in phase-based
networks, not all neurons fire in each period of the oscillation. This suggests that neurons
may employ a dual code ? the more rate-based probability of being active in a cycle, and
the phase-based timing of the spike relative to the cycle [24]. The advantages of such a
scheme have yet to be fully characterized.
Second, in the present framework the choice of the learning rule is arbitrary, as long as the
A
0.35
0.3
B
prior
input
Bayesian
6
0.25
5
Average error
Frequency
7
0.2
0.15
4
3
0.1
2
0.05
1
0
?60
?40
?20
0
Error
20
40
60
input
ideal observer
Bayesian
0
1
10
Number of stored patterns
100
Figure 2: Performance of the phase-coded network. Error distribution for the ideal observer was
a Dirac-delta at 0 (B) and was thus omitted from A. Average error of the ?prior only? network was
too large (A) to be plotted in B. The prior was a von Mises distribution with ?x = 0 mean, ?x =
0.5 concentration on a T? = 125 ms long cycle matching data on theta frequency modulation
of pyramidal cell population activity in the hippocampus [23]. Input was corrupted by unbiased
circular Gaussian (von Mises) noise with ?x = 10 concentration. Learning rule was circular STDP
rule with ASTDP = 0.03, ?STDP = 4 and TSTDP = T? parameters matching experimental data on
hippocampal STDP [5] and theta periodicity. The network consisted of N = 100 cells, and the
number of memories stored was M = 10 (A) or varied between M = 2 . . . 100 (B, note logarithmic
scale). For further explanation of symbols and axes, see Figure 1.
recall dynamics is optimally matched to it. Our formalism also suggests that there may be
a way to optimally choose the learning rule itself in the first place, by matching it to the
prior distribution of patterns. This approach would thus be fundamentally different from
those seeking ?globally? optimal learning rules [25], and may be more similar to those used
to find optimal tuning curves appropriately matching stimulus statistics [26].
References
[1]
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
[21]
[22]
[23]
[24]
[25]
[26]
Marr D. Philos Trans R Soc Lond B Biol Sci 262:23, 1971.
O?Keefe J. Exp Neurol 51:78, 1976.
O?Keefe J, Recce ML. Hippocampus 3:317, 1993.
Bliss TVP, L?mo T. J Physiol (Lond) 232:331, 1973.
Bi GQ, Poo MM. J Neurosci 18:10464, 1998.
MacKay DJC. In Maximum entropy and Bayesian methods, 237, 1990.
Sommer FT, Dayan P. IEEE Trans Neural Netw 9:705, 1998.
Jaynes ET. Probability theory: the logic of science. Cambridge University Press, 2003.
Amit DJ. Modeling brain function. Cambridge University Press, 1989.
Dayan P, Abbott LF. Theoretical neuroscience. MIT Press, 2001.
Zhang K. J Neurosci 16:2112, 1996.
Seung HS. Proc Natl Acad Sci USA 93:13339, 1996.
Treves A. Phys Rev A 42:2418, 1990.
Hopfield JJ. Proc Natl Acad Sci USA 76:2554, 1982.
Hopfield JJ. Proc Natl Acad Sci USA 81:3088, 1984.
Buzs?ki Gy. Neuron 33:325, 2002.
Harris KD, et al. Nature 424:552, 2003.
Li Z, Hopfield JJ. Biol Cybern 61:379, 1989.
Abbott LF, Nelson SB. Nat Neurosci 3:1178, 2000.
Scarpetta S, et al. Neural Comput 14:2371, 2002.
Ermentrout B, et al. Neural Comput 13:1285, 2001.
Reyes AD, Fetz FE. J Neurophysiol 69:1673, 1993.
Klausberger T, et al. Nature 421:844, 2003.
Mueller R, et al. In BioNet?96 , 70, 1976.
Gardner E, Derrida B. J Phys A 21:271, 1988.
Laughlin S. Z Naturforsch 36:901, 1981.
| 2683 |@word h:1 version:1 compression:1 seems:1 hippocampus:4 simulation:3 crucially:1 covariance:4 thereby:1 reduction:1 united:1 tuned:2 current:1 jaynes:1 yet:1 dx:1 import:1 physiol:1 numerical:2 additive:2 distant:1 plasticity:12 shape:1 designed:1 update:5 characterization:1 provides:1 contribute:1 simpler:1 zhang:1 ipx:3 specialize:1 manner:1 expected:1 brain:4 integrator:1 touchstone:1 globally:1 actual:1 increasing:1 provided:1 matched:2 benchmarked:1 interpreted:2 finding:1 biphasic:1 questionable:1 exactly:2 uk:1 tricky:1 unit:4 timing:6 treat:2 local:6 limit:1 acad:3 despite:2 encoding:1 firing:15 modulation:1 approximately:3 might:1 plus:1 chose:1 suggests:3 autoassociation:1 co:3 range:2 bi:1 ternary:5 lf:2 episodic:1 area:4 gabor:1 matching:4 imprecise:1 pre:5 regular:1 suggest:1 get:1 close:2 storage:2 context:1 cybern:1 optimize:2 equivalent:1 map:3 conventional:1 poo:1 straightforward:1 rule:24 insight:1 regarded:1 importantly:1 marr:1 population:1 updated:2 construction:2 exact:2 substrate:1 element:2 cut:1 observed:1 ft:1 solved:1 unexpectedly:1 wj:8 region:2 connected:1 cycle:3 decrease:1 environment:1 leak:1 complexity:1 moderately:1 seung:1 ermentrout:1 dynamic:24 neglected:1 depend:2 purely:1 neurophysiol:1 hopfield:3 various:3 represented:1 derivation:3 effective:1 london:2 precedes:1 quite:2 apparent:1 whose:1 plausible:1 encoded:2 widely:1 solve:1 otherwise:1 cov:1 statistic:1 noisy:4 itself:2 advantage:1 ucl:1 interaction:10 interconnected:1 reset:3 gq:1 relevant:1 combining:1 fired:1 description:1 validate:1 dirac:3 requirement:1 boris:1 perfect:1 wider:1 derive:3 depending:2 ac:1 pose:1 coupling:1 derrida:1 eq:6 soc:1 implemented:1 involves:4 indicate:1 come:1 differ:1 direction:2 correct:1 centered:1 require:1 stayed:1 biological:5 hold:1 practically:1 around:1 sufficiently:1 considered:2 stdp:14 exp:3 mm:1 mapping:1 mo:1 a2:1 omitted:1 proc:3 outperformed:1 gauge:1 autapses:1 weighted:2 mit:1 clearly:1 gaussian:12 rather:3 probabilistically:1 ax:1 improvement:1 likelihood:1 indicates:1 greatly:1 contrast:1 baseline:2 helpful:1 inference:5 posteriori:1 dayan:4 dependent:6 sin2:1 mueller:1 sb:1 issue:1 among:1 ill:1 dual:1 animal:1 resonance:1 special:1 integration:1 mackay:1 field:4 having:4 future:1 report:1 stimulus:2 fundamentally:1 employ:1 randomly:2 individual:1 phase:17 fire:2 attractor:2 attempt:1 recalling:1 conductance:2 interest:1 regulating:1 djc:1 possibility:1 circular:4 highly:1 adjust:1 saturated:1 arrives:1 yielding:1 light:2 natl:3 wc1n:1 encourage:1 partial:1 necessary:2 unless:1 initialized:1 plotted:2 theoretical:3 instance:1 formalism:2 modeling:1 ar:1 measuring:1 queen:1 ca3:2 rare:1 delay:1 too:1 optimally:2 stored:25 corrupted:4 considerably:1 density:3 peak:1 probabilistic:2 decoding:1 together:1 connecting:1 squared:1 again:3 von:2 opposed:1 choose:1 li:1 account:3 potential:2 gy:1 coding:1 includes:2 coefficient:1 bliss:1 satisfy:1 explicitly:1 caused:1 depends:2 ad:1 later:2 root:2 performed:1 observer:13 bayes:1 decaying:1 contribution:1 ass:1 square:2 variance:10 characteristic:6 resetting:1 largely:1 sitting:1 generalize:1 bayesian:21 biophysically:1 iid:1 lengyel:1 j6:5 oscillatory:1 synapsis:3 reach:2 phys:2 synaptic:33 against:3 frequency:5 involved:3 pp:1 transmits:2 dxi:2 mi:2 couple:1 gain:1 recall:16 knowledge:2 carefully:2 back:2 factorizability:1 dt:1 specify:3 synapse:1 though:1 governing:1 just:2 hand:1 lack:1 quality:1 perhaps:1 lossy:1 usa:3 effect:2 requiring:1 unbiased:5 normalized:2 true:1 former:1 equality:1 consisted:1 symmetric:4 sin:1 during:2 self:3 encourages:1 lumped:1 essence:1 suboptimality:1 m:1 hippocampal:4 prominent:2 complete:3 demonstrate:1 neocortical:1 performs:2 reyes:1 reflection:1 consideration:1 superior:1 spiking:2 exponentially:1 extend:1 approximates:1 interpret:1 expressing:1 cambridge:2 tuning:1 philos:1 hp:2 dj:1 gutkin:1 specification:2 access:1 longer:1 surface:1 inhibition:3 buzs:1 posterior:3 perspective:1 retrieved:1 binary:3 came:1 captured:1 greater:1 determine:1 maximize:1 period:1 signal:1 ii:1 full:2 sound:1 hebbian:4 match:2 characterized:2 offer:1 long:3 retrieval:1 post:2 coded:7 controlled:1 involving:1 basic:1 histogram:1 cell:15 wealth:1 wake:1 pyramidal:1 appropriately:1 unlike:2 finely:1 ascent:2 fell:1 effectiveness:1 seem:1 extracting:1 near:1 ideal:14 variety:2 xj:36 perfectly:2 imperfect:1 simplifies:1 effort:1 render:1 peter:1 constitute:1 jj:3 autoassociative:12 ignored:1 generally:1 useful:1 clear:2 simplest:1 specifies:1 canonical:6 sign:1 neuroscience:2 arising:1 delta:3 correctly:1 bulk:1 write:1 discrete:1 express:2 key:1 nevertheless:1 threshold:2 drawn:5 penalizing:1 abbott:2 lmate:1 sum:4 angle:1 uncertainty:2 place:2 laid:1 reasonable:3 oscillation:6 decision:1 comparable:1 accelerates:2 bound:3 hi:3 ki:1 sleep:1 activity:30 constraint:2 constrain:1 x2:11 generates:2 aspect:2 speed:1 lond:2 px:3 kd:1 membrane:2 describes:3 postsynaptic:10 wi:20 rev:1 biologically:5 making:1 restricted:1 computationally:2 equation:9 know:1 whichever:1 end:1 permit:2 appropriate:5 alternative:1 odor:1 original:2 remaining:1 sommer:1 coni:1 embodies:1 amit:1 graded:8 leakage:1 implied:1 objective:4 seeking:1 quantity:1 spike:10 concentration:2 unclear:1 exhibit:1 gradient:2 thank:1 link:1 simulated:1 capacity:1 sci:4 nelson:1 presynaptic:4 extent:1 trivial:1 assuming:1 code:1 relationship:1 exquisitely:1 ratio:1 kingdom:1 unfortunately:1 troubling:1 tribute:1 fe:1 stated:2 negative:2 implementation:1 perform:1 upper:2 neuron:27 anti:2 truncated:4 situation:1 incorporated:1 varied:2 arbitrary:1 treves:11 specified:1 extensive:1 optimized:1 recalled:4 coherent:2 narrow:1 acov:6 trans:2 address:1 bar:1 dynamical:1 pattern:45 xm:5 below:1 regime:3 reading:1 challenge:1 saturation:1 including:3 memory:23 explanation:1 event:1 natural:2 restore:1 recce:1 scheme:1 theta:4 gardner:1 axis:2 coupled:2 extract:1 prior:32 understanding:1 discovery:1 determining:2 relative:8 fully:1 brittle:1 interesting:1 suggestion:1 proportional:2 foundation:1 degree:1 offered:1 charitable:1 corrupting:3 storing:2 systematically:1 course:2 summary:1 periodicity:1 supported:1 last:2 asynchronous:1 truncation:1 side:1 allow:1 bias:2 laughlin:1 fetz:1 wide:1 fall:1 taking:1 face:1 emerge:1 leaky:1 sparse:1 curve:4 calculated:2 cortical:1 computes:1 concretely:1 made:1 clue:1 founded:1 approximate:2 netw:1 neurobiological:1 logic:1 ml:1 global:1 active:1 assumed:2 factorize:1 xi:67 scarpetta:1 search:1 iterative:1 quantifies:1 naturforsch:1 nature:2 reasonably:1 inherently:1 ignoring:1 astdp:3 argmaxx:1 necessarily:1 antisymmetric:1 did:4 neurosci:3 noise:12 arise:1 competent:1 formul:1 convey:1 x1:5 neuronal:16 fig:9 referred:1 benchmarking:1 gatsby:3 position:1 comput:2 lie:1 third:3 formula:1 showing:1 recurrently:2 list:2 decay:4 symbol:1 neurol:1 normalizing:1 evidence:2 essential:1 a3:1 adding:3 effectively:2 importance:1 keefe:2 magnitude:2 nat:1 horizon:1 entropy:1 logarithmic:2 simply:1 likely:1 determines:1 satisfies:1 harris:1 conditional:2 identity:1 absence:1 change:3 determined:5 except:1 acting:1 experimental:2 meaningful:2 formally:1 college:1 support:2 latter:1 arises:2 assessed:1 ongoing:1 tested:1 biol:2 |
1,854 | 2,684 | Joint MRI Bias Removal Using Entropy
Minimization Across Images
Erik G. Learned-Miller
Department of Computer Science
University of Massachusetts, Amherst
Amherst, MA 01003
Parvez Ahammad
Division of Electrical Engineering
University of California, Berkeley
Berkeley, CA 94720
Abstract
The correction of bias in magnetic resonance images is an important
problem in medical image processing. Most previous approaches have
used a maximum likelihood method to increase the likelihood of the pixels in a single image by adaptively estimating a correction to the unknown
image bias field. The pixel likelihoods are defined either in terms of a
pre-existing tissue model, or non-parametrically in terms of the image?s
own pixel values. In both cases, the specific location of a pixel in the image is not used to calculate the likelihoods. We suggest a new approach
in which we simultaneously eliminate the bias from a set of images of
the same anatomy, but from different patients. We use the statistics from
the same location across different images, rather than within an image, to
eliminate bias fields from all of the images simultaneously. The method
builds a ?multi-resolution? non-parametric tissue model conditioned on
image location while eliminating the bias fields associated with the original image set. We present experiments on both synthetic and real MR
data sets, and present comparisons with other methods.
1
Introduction
The problem of bias fields in magnetic resonance (MR) images is an important problem
in medical imaging. This problem is illustrated in Figure 1. When a patient is imaged in
the MR scanner, the goal is to obtain an image which is a function solely of the underlying
tissue (left of Figure 1). However, typically the desired anatomical image is corrupted by a
multiplicative bias field (2nd image of Figure 1) that is caused by engineering issues such
as imperfections in the radio frequency coils used to record the MR signal. The result is a
corrupted image (3rd image of Figure 1). (See [1] for background information.) The goal
of MR bias correction is to estimate the uncorrupted image from the corrupted image.
A variety of statistical methods have been proposed to address this problem. Wells et
al. [7] developed a statistical model using a discrete set of tissues, with the brightness
distribution for each tissue type (in a bias-free image) represented by a one-dimensional
Guassian distribution. An expectation-maximization (EM) procedure was then used to
simultaneouly estimate the bias field, the tissue type, and the residual noise. While this
method works well in many cases, it has several drawbacks: (1) Models must be developed
a priori for each type of acquistion (for each different setting of the MR scanner), for each
Figure 1: On the left is an idealized mid-axial MR image of the human brain with little
or no bias field. The second image is a simulated low-frequency bias field. It has been
exaggerated for ease of viewing. The third image is the result of pixelwise multiplication
of the image by the bias field. The goal of MR bias correction is to recover the low-bias
image on the left from the biased image on the right. On the right is the sine/cosine basis,
used to construct band-limited bias fiels (see text).
new area of the body, and for different patient populations (like infants and adults). (2)
Models must be developed from ?bias-free? images, which may be difficult or impossible
to obtain in many cases. (3) The model assumes a fixed number of tissues, which may
be inaccurate. For example, during development of the human brain, there is continuous
variability between gray matter and white matter. In addition, a discrete tissue model does
not handle so-called partial volume effects in which a pixel represents a combination of
several tissue types. This occurs frequently since many pixels occur at tissue boundaries.
Non-parametric approaches have also been suggested, as for example by Viola [10]. In
that work, a non-parametric model of the tissue was developed from a single image. Using
the observation that the entropy of the pixel brightness distribution for a single image is
likely to increase when a bias field is added, Viola?s method postulates a bias-correction
field by minimizing the entropy of the resulting pixel brightness distribution. This approach
addresses several of the problems of fixed-tissue parametric models, but has its own drawbacks: (1) The statistical model may be weak, since it is based on data from only a single
image. (2) There is no mechanism for distinguishing between certain low-frequency image
components and a bias field. That is, the method may mistake signal for noise in certain
cases when removal of the true signal reduces the entropy of the brightness distriibution.
We shall show that this is a problem in real medical images.
The method we present overcomes or improves upon problems associated with both of
these methods and their many variations (see, e.g., [1] for recent techniques). It models tissue brightness non-parametrically, but uses data from multiple images to provide improved
distribution estimates and alleviate the need for bias-free images for making a model. It
also conditions on spatial location, taking advantage of a rich information source ignored
in other methods. Experimental results demonstrate the effectiveness of our method.
2
The Image Model and Problem Formulation
We assume we are given a set I of observed images Ii with 1 ? i ? N, as shown on the
left side of Figure 2. Each of these images is assumed to be the product of some bias-free
image Li and a smooth bias field Bi ? B . We shall refer to the bias-free images as latent
images (also called intrinsic images by some authors). The set of all latent images shall be
denoted L and the set of unknown bias fields B. Then each observed image can be written
as the product Ii (x, y) = Li (x, y) ? Bi (x, y), where (x, y) gives the pixel coordinates of each
point, with P pixels per image.
Consider again Figure 2. A pixel-stack through each image set is shown as the set of pixels
corresponding to a particular location in each image (not necessarily the same tissue type).
Our method relies on the principle that the pixel-stack values will have lower entropy when
the bias fields have been removed. Figure 3 shows the simulated effect, on the distribution
of values in a pixel-stack, of adding different bias fields to each image.
The latent image generation model assumes that each pixel is drawn from a fixed distribution px,y (?) which gives the probability of each gray value at the the location (x, y) in the
image. Furthermore, we assume that all pixels in the latent image are independent, given
the distributions from which they are drawn. It is also assumed that the bias fields for each
image are chosen independently from some fixed distribution over bias fields. Unlike most
models for this problem which rely on statistical regularities within an image, we take a
completely orthogonal approach by assuming that pixel values are independent given their
image locations, but that pixel-stacks in general have low entropy when bias fields are
removed.
We formulate the problem as a maximum a posteriori (MAP) problem, searching for the
most probable bias fields given the set of observed images. Letting B represent the 25dimensional product space of smooth bias fields (corresponding to the 25 basis images of
Figure 1), we wish to find
arg maxP(B|I)
(a)
=
B?B
arg maxP(I|B)P(B)
(1)
B?B
(b)
=
arg maxP(I|B)
(2)
B?B
(c)
=
arg maxP(L(I, B))
(3)
B?B
N
=
arg max ? ? px,y (Li (x, y))
B?B
(4)
x,y i=1
N
=
arg max ? ? log px,y (Li (x, y))
(5)
arg min ? H(px,y )
(6)
arg min ? H? Vasicek (L1 (x, y), ..., LN (x, y))
(7)
B?B
(d)
?
B?B
(e)
?
B?B
=
x,y i=1
x,y
x,y
arg min ? H? Vasicek (
B?B
x,y
I1 (x, y)
IN (x, y)
, ...,
).
B1 (x, y)
BN (x, y)
(8)
Here H is the Shannon entropy (?E(log P(x))) and H? Vasicek is a sample-based entropy
estimator.1 (a) is just an application of Bayes rule. (b) assumes a uniform prior over the
allowed bias fields. The method can easily be altered to incorporate a non-uniform prior.
1 The
entropy estimator used is similar to Vasicek?s estimator [6], given (up to minor details) by
N (i+m)
1 N?m
(i)
H? Vasicek (Z 1 , ..., Z N ) =
log
(Z
?
Z
)
,
(9)
?
N ? m i=1
m
where Z i ?s represent the values in a pixel-stack, Z (i) ?s represent those same values in rank order, N is
the number of values in the pixel-stack and m is a function of N (like N 0.5 ) such that m/N goes to 0
as m and N go to infinity. These entropy estimators are discussed at length elsewhere [3].
Figure 2: On the left are a set of mid-coronal brain images from eight different infants,
showing clear signs of bias fields. A pixel-stack, a collection of pixels at the same point
in each image, is represented by the small square near the top of each image. Although
there are probably no more than two or three tissue types represented by the pixel-stack,
the brightness distribution through the pixel-stack has high empirical entropy due to the
presence of different bias fields in each image. On the right are a set of images that have
been corrected using our bias field removal algorithm. While the images are still far from
identical, the pixel-stack entropies have been reduced by mapping similar tissues to similar
values in an ?unsupervised? fashion, i.e. without knowing or estimating the tissue types.
(c) expresses the fact that the probability of the observed image given a particular bias field
is the same as the probability of the latent image associated with that observed image and
bias field. The approximation (d) replaces the empirical mean of the log probability at each
pixel with the negative entropy of the underlying distribution at that pixel. This entropy is
in turn estimated (e) using the entropy estimator of Vasicek [6] directly from the samples in
the pixel-stack, without ever estimating the distributions px,y explicitly. The inequality (d)
becomes an equality as N grows large by the law of large numbers, while the consistency
of Vasicek?s entropy estimator [2] implies that (e) also goes to equality with large N. (See
[2] for a review of entropy estimators.)
3
The Algorithm
Using these ideas, it is straightforward to construct algorithms for joint bias field removal.
As mentioned above, we chose to optimize Equation (8) over the set of band-limited bias
fields. To do this, we parameterize the set of bias fields using the sine/cosine basis images
shown on the right of Figure 1:
25
Bi =
? ? j ? j (x, y).
j=1
We optimize Equation (8) by simultaneously updating the bias field estimates (taking a step
along the numerical gradient) for each image to reduce the overall entropy. That is, at time
step t, the coefficients ? j for each bias field are updated using the latent image estimates
and entropy estimates from time step t ? 1. After all ??s have been updated, a new set of
latent images and pixel-stack entropies are calculated, and another gradient step is taken.
Though it is possible to do a full gradient descent to convergence by optimizing one image
at a time, the optimization landscape tends to have more local minima for the last few
images in the process. The appeal of our joint gradient descent method, on the other hand,
is that the ensemble of images provides a natural smoothing of the optimization landscape
8
8
7
7
6
6
5
5
4
4
3
3
2
2
1
0
1
0
50
100
150
200
250
0
0
50
100
150
200
250
Figure 3: On the left is a simulated distribution from a pixel-stack taken through a particular set of bias-free mid-axial MR images. The two sharp peaks in the brightness distribution
represent two tissues which are commonly found at that particular pixel location. On the
right is the result of adding an independent bias field to each image. In particular, the
spread, or entropy, of the pixel distribution increases. In this work, we seek to remove bias
fields by seeking to reduce the entropy of the pixel-stack distribution to its original state.
in the joint process. It is in this sense that our method is ?multi-resolution?, proceeding
from a smooth optimization in the beginning to a sharper one near the end of the process.
We now summarize the algorithm:
1. Initialize the bias field coefficients for each image to 0, with the exception of
the coefficient for the DC-offset (the constant bias field component), which is
initialized to 1. Initialize the gradient descent step size ? to some value.
2. Compute the summed pixelwise entropies for the set of images with initial ?neutral? bias field corrections. (See below for method of computation.)
3. Iterate the following loop until no further changes occur in the images.
(a) For each image:
i. Calculate the numerical gradient ?? HVasicek of (8) with respect to the bias
field coefficients (? j ?s) for the current image.
ii. Set ? = ? + ??? H? Vasicek .
(b) Update ? (reduce its value according to some schedule).
Upon convergence, it is assumed that the entropy has been reduced as much as possible by
changing the bias fields, unless one or more of the gradient descents is stuck in a local minimum. Empirically, the likelihood of sticking in local minima is dramatically reduced by
increasing the number of images (N) in the optimization. In our experiments described below with only 21 real infant brains, the algorithm appears to have found a global minimum
of all bias fields, at least to the extent that this can be discerned visually.
Note that for a set of identical images, the pixel-stack entropies are not increased by multiplying each image by the same bias field (since all images will still be the same). More
generally, when images are approximately equivalent, their pixel-stack entropies are not
signficantly affected by a ?common? bias field, i.e. one that occurs in all of the images.2
This means that the algorithm cannot, in general, eliminate all bias fields from a set of images, but can only set all of the bias fields to be equivalent. We refer to any constant bias
field remaining in all of the images after convergence as the residual bias field.
2 Actually, multiplying each image by a bias field of small magnitude can artificially reduce the
entropy of a pixel-stack, but this is only the result of the brightness values shrinking towards zero.
Such artificial reductions in entropy can be avoided by normalizing a distribution to unit variance
between iterations of computing its entropy, as is done in this work.
Fortunately, there is an effect that tends to minimize the impact of the residual bias field
in many test cases. In particular, the residual bias field tends to consist of components for
each ? j that approximate the mean of that component across images. For example, if half
of the observed images have a positive value for a particular component?s coefficient, and
half have a negative coefficient for that component, the residual bias field will tend to have
a coefficient near zero for that component. Hence, the algorithm naturally eliminates bias
field effects that are non-systematic, i.e. that are not shared across images.
If the same type of bias field component occurs in a majority of the images, then the algorithm will not remove it, as the component is indistinguishable, under our model, from the
underlying anatomy. In such a case, one could resort to within-image methods to further
reduce the entropy. However, there is a risk that such methods will remove components
that actually represent smooth gradations in the anatomy. This can be seen in the bottom
third of Figure 4, and will be discussed in more detail below.
4
Experiments
To test our algorithm, we ran two sets of experiments, the first on synthetic images for
validation, and the second on real brain images. We obtained synthetic brain images from
the BrainWeb project [8, 9] such as the one shown on the left of Figure 1. These images can
be considered ?idealized? MR images in the sense that the brightness values for each tissue
are constant (up to a small amount of manually added isotropic noise). That is, they contain
no bias fields. The initial goal was to ensure that our algorithm could remove synthetically
added bias fields, in which the bias field coefficients were known. Using K copies of a
single ?latent? image, we added known but different bias fields to each one. For as few as
five images, we could reliably recover the known bias field coefficients, up to a fixed offset
for each image, to within 1% of the power of the original bias coefficients.
More interesting are the results on real images, in which the latent images come from
different patients. We obtained 21 pre-registered3 infant brain images (top of Figure 4)
from Brigham and Women?s Hospital in Boston, Massachusetts. Large bias fields can be
seen in many of the images. Probably the most striking is a ?ramp-like? bias field in the
sixth image of the second row. (The top of the brain is too bright, while the bottom is too
dark.) Because the brain?s white matter is not fully developed in these infant scans, it is
difficult to categorize tissues into a fixed number of classes as is typically done for adult
brain images; hence, these images are not amenable to methods based on specific tissue
models developed for adults (e.g. [7]).
The middle third of Figure 4 shows the results of our algorithm on the infant brain images.
(These results must be viewed in color on a good monitor to fully appreciate the results.)
While a trained technician can see small imperfections in these images, the results are
remarkably good. All major bias artifacts have been removed.
It is interesting to compare these results to a method that reduces the entropy of each image
individually, without using constraints between images. Using the results of our algorithm
as a starting point, we continued to reduce the entropy of the pixels within each image
(using a method akin to Viola?s [10]), rather than across images. These results are shown
in the bottom third of Figure 4. Carefully comparing the central brain regions in the middle
section of the figure and the bottom section of the figure, one can see that the butterfly
shaped region in the middle of the brain, which represents developing white matter, has
3 It
is interesting to note that registration is not strictly necessary for this algorithm to work. The
proposed MAP method works under very broad conditions, the main condition being that the bias
fields do not span the same space as parts of the actual medical images. It is true, however, that as the
latent images become less registered or differ in other ways, that a much larger number of images is
needed to get good estimates of the pixel-stack distributions.
been suppressed in the lower images. This is most likely because the entropy of the pixels
within a particular image can be reduced by increasing the bias field ?correction? in the
central part of the image. In other words, the algorithm strives to make the image more
uniform by removing the bright part in the middle of the image. However, our algorithm,
which compares pixels across images, does not suppress these real structures, since they
occur across images. Hence coupling across images can produce superior results.
5
Discussion
The idea of minimizing pixelwise entropies to remove nuisance variables from a set of images is not new. In particular, Miller et al. [4, 5] presented an approach they call congealing
in which the sum of pixelwise entropies is minimized by separate affine transforms applied
to each image. Our method can thus be considered an extension of the congealing process
to non-spatial transformations. Combining such approaches to do registration and bias removal simulataneously, or registration and lighting rectification of faces, for example, is an
obvious direction for future work.
This work uses information unused in other methods, i.e. information across images. This
suggests an iterative scheme in which both types of information, both within and across
images, are used. Local models could be based on weighted neighborhoods of pixels, pixel
cylinders, rather than single pixel-stacks, in sparse data scenarios. For ?easy? bias correction problems, such an approach may be overkill, but for difficult problems in bias correction, where the bias field is difficult to separate from the underlying tissue, as discussed in
[1], such an approach could produce critical extra leverage.
We would like to thank Dr. Terrie Inder and Dr. Simon Warfield for graciously providing
the infant brain images for this work. The images were obtained under NIH grant P41
RR13218. Also, we thank Neil Weisenfeld and Sandy Wells for helpful discussions. This
work was partially supported by Army Research Office grant DAAD 19-02-1-0383.
References
[1] Fan, A., Wells, W., Fisher, J., Cetin, M., Haker, S., Mulkern, C., Tempany, C., Willsky, A. A
unified variational approach to denoising and bias correction in MR. Proceedings of IPMI, 2003.
[2] Beirlant, J., Dudewicz, E., Gyorfi, L. and van der Meulen, E. Nonparametric entropy estimation:
An overview. International Journal of Mathematical and Statistical Sciences, 6. pp.17-39. 1997.
[3] Learned-Miller, E. G. and Fisher, J. ICA using spacings estimates of entropy. Journal of Machine
Learning Research, Volume 4, pp. 1271-1295, 2003.
[4] Miller, E. G., Matsakis, N., Viola, P. A. Learning from one example through shared densities on
transforms. IEEE Conference on Computer Vision and Pattern Recognition. 2000.
[5] Miller, E. G. Learning from one example in machine vision by sharing probability densities.
Ph.D. thesis. Massachusetts Institute of Technology. 2002.
[6] Vasicek, O. A test for normality based on sample entropy. Journal of the Royal Statistical Society
Series B, 31. pp. 632-636, 1976.
[7] Wells, W. M., Grimson, W. E. L., Kikinis, R., Jolesz, F. Adaptive segmentation of MRI data.
IEEE Transactions on Medical Imaging, 15. pp. 429-442, 1996.
[8] Collins, D.L., Zijdenbos, A.P., Kollokian, J.G., Sled, N.J., Kabani, C.J., Holmes, C.J., Evans,
A.C. Design and Construction of a realistic digital brain phantom. IEEE Transactions on Medical
Imaging, 17. pp. 463-468, 1998.
[9] http://www.bic.mni.mcgill.ca/brainweb/
[10] Viola, P.A. Alignment by maximization of mutual information. Ph.D. Thesis. Massachusetts
Institute of Technology. 1995.
Figure 4: NOTE: This image must be viewed in color (preferably on a bright display) for
full effect. Top. Original infant brain images. Middle. The same images after bias removal
with our algorithm. Note that developing white matter (butterfly-like structures in middle
brain) is well-preserved. Bottom. Bias removal using a single image based algorithm.
Notice that white matter structures are repressed.
| 2684 |@word middle:6 mri:2 eliminating:1 nd:1 seek:1 bn:1 brightness:9 reduction:1 initial:2 series:1 existing:1 current:1 comparing:1 must:4 written:1 evans:1 realistic:1 numerical:2 remove:5 update:1 infant:8 half:2 isotropic:1 beginning:1 record:1 provides:1 location:8 five:1 mathematical:1 along:1 become:1 ica:1 frequently:1 multi:2 brain:17 little:1 actual:1 increasing:2 becomes:1 project:1 estimating:3 underlying:4 developed:6 unified:1 transformation:1 berkeley:2 preferably:1 unit:1 medical:6 grant:2 positive:1 cetin:1 engineering:2 local:4 tends:3 mistake:1 solely:1 approximately:1 chose:1 suggests:1 ease:1 limited:2 bi:3 gyorfi:1 procedure:1 area:1 empirical:2 pre:2 word:1 suggest:1 get:1 cannot:1 risk:1 impossible:1 gradation:1 www:1 optimize:2 equivalent:2 map:2 phantom:1 go:3 straightforward:1 starting:1 independently:1 resolution:2 formulate:1 estimator:7 rule:1 continued:1 holmes:1 population:1 handle:1 searching:1 variation:1 coordinate:1 updated:2 mcgill:1 construction:1 distinguishing:1 us:2 recognition:1 updating:1 observed:6 bottom:5 electrical:1 parameterize:1 calculate:2 region:2 removed:3 ran:1 mentioned:1 grimson:1 ipmi:1 trained:1 upon:2 division:1 basis:3 completely:1 easily:1 joint:4 represented:3 guassian:1 coronal:1 artificial:1 congealing:2 neighborhood:1 larger:1 ramp:1 maxp:4 statistic:1 neil:1 butterfly:2 advantage:1 product:3 loop:1 combining:1 sticking:1 convergence:3 regularity:1 produce:2 coupling:1 axial:2 minor:1 implies:1 come:1 differ:1 direction:1 anatomy:3 drawback:2 human:2 viewing:1 alleviate:1 probable:1 strictly:1 extension:1 correction:10 scanner:2 considered:2 visually:1 mapping:1 major:1 sandy:1 estimation:1 radio:1 individually:1 weighted:1 minimization:1 imperfection:2 rather:3 office:1 rank:1 likelihood:5 sense:2 graciously:1 posteriori:1 helpful:1 inaccurate:1 eliminate:3 typically:2 i1:1 pixel:43 issue:1 arg:9 overall:1 denoted:1 priori:1 development:1 resonance:2 spatial:2 smoothing:1 initialize:2 summed:1 mutual:1 field:62 construct:2 shaped:1 manually:1 identical:2 represents:2 broad:1 unsupervised:1 future:1 minimized:1 few:2 simultaneously:3 cylinder:1 alignment:1 amenable:1 partial:1 necessary:1 orthogonal:1 unless:1 initialized:1 desired:1 vasicek:9 acquistion:1 increased:1 maximization:2 parametrically:2 neutral:1 uniform:3 too:2 pixelwise:4 corrupted:3 synthetic:3 adaptively:1 density:2 peak:1 amherst:2 international:1 systematic:1 again:1 postulate:1 central:2 thesis:2 woman:1 dr:2 resort:1 li:4 coefficient:10 matter:6 caused:1 explicitly:1 idealized:2 multiplicative:1 sine:2 recover:2 bayes:1 simon:1 minimize:1 square:1 bright:3 variance:1 miller:5 ensemble:1 landscape:2 weak:1 daad:1 multiplying:2 lighting:1 tissue:22 sharing:1 sixth:1 frequency:3 pp:5 obvious:1 naturally:1 associated:3 massachusetts:4 color:2 improves:1 segmentation:1 schedule:1 carefully:1 actually:2 appears:1 improved:1 discerned:1 formulation:1 done:2 though:1 furthermore:1 just:1 until:1 hand:1 artifact:1 gray:2 grows:1 effect:5 contain:1 true:2 equality:2 hence:3 imaged:1 illustrated:1 white:5 indistinguishable:1 during:1 nuisance:1 cosine:2 demonstrate:1 l1:1 image:136 variational:1 nih:1 common:1 superior:1 empirically:1 overview:1 volume:2 discussed:3 refer:2 rd:1 consistency:1 own:2 recent:1 exaggerated:1 optimizing:1 scenario:1 certain:2 inequality:1 der:1 uncorrupted:1 seen:2 minimum:4 fortunately:1 mr:11 brainweb:2 signal:3 ii:3 multiple:1 full:2 reduces:2 smooth:4 technician:1 impact:1 dudewicz:1 patient:4 expectation:1 vision:2 iteration:1 represent:5 preserved:1 background:1 addition:1 remarkably:1 spacing:1 source:1 biased:1 eliminates:1 unlike:1 extra:1 probably:2 tend:1 effectiveness:1 call:1 near:3 presence:1 synthetically:1 unused:1 leverage:1 easy:1 variety:1 iterate:1 bic:1 reduce:6 idea:2 knowing:1 inder:1 akin:1 ignored:1 dramatically:1 generally:1 clear:1 amount:1 transforms:2 dark:1 mid:3 band:2 nonparametric:1 ph:2 reduced:4 http:1 notice:1 sign:1 estimated:1 per:1 anatomical:1 discrete:2 shall:3 affected:1 express:1 monitor:1 drawn:2 changing:1 registration:3 imaging:3 sum:1 striking:1 display:1 fan:1 replaces:1 mni:1 rr13218:1 occur:3 infinity:1 constraint:1 min:3 span:1 px:5 department:1 signficantly:1 according:1 developing:2 combination:1 across:10 strives:1 em:1 suppressed:1 jolesz:1 making:1 taken:2 ln:1 equation:2 rectification:1 turn:1 mechanism:1 needed:1 letting:1 end:1 eight:1 magnetic:2 matsakis:1 original:4 assumes:3 top:4 remaining:1 ensure:1 build:1 society:1 appreciate:1 seeking:1 added:4 occurs:3 parametric:4 gradient:7 separate:2 thank:2 simulated:3 majority:1 extent:1 willsky:1 assuming:1 erik:1 length:1 providing:1 minimizing:2 difficult:4 sharper:1 negative:2 suppress:1 design:1 reliably:1 unknown:2 observation:1 descent:4 viola:5 variability:1 ever:1 dc:1 stack:19 sharp:1 california:1 learned:2 registered:1 address:2 adult:3 suggested:1 below:3 pattern:1 summarize:1 max:2 royal:1 power:1 critical:1 natural:1 rely:1 residual:5 normality:1 scheme:1 altered:1 meulen:1 technology:2 text:1 prior:2 review:1 removal:7 multiplication:1 law:1 fully:2 kikinis:1 generation:1 interesting:3 validation:1 digital:1 affine:1 principle:1 row:1 elsewhere:1 supported:1 last:1 free:6 copy:1 bias:83 side:1 institute:2 taking:2 face:1 overkill:1 sparse:1 van:1 boundary:1 calculated:1 rich:1 author:1 collection:1 commonly:1 stuck:1 avoided:1 adaptive:1 far:1 transaction:2 approximate:1 overcomes:1 global:1 b1:1 assumed:3 continuous:1 latent:10 iterative:1 ca:2 necessarily:1 artificially:1 beirlant:1 spread:1 main:1 noise:3 allowed:1 body:1 fashion:1 shrinking:1 wish:1 third:4 removing:1 specific:2 showing:1 appeal:1 offset:2 normalizing:1 brigham:1 intrinsic:1 consist:1 adding:2 magnitude:1 p41:1 conditioned:1 boston:1 entropy:38 likely:2 army:1 partially:1 relies:1 ma:1 coil:1 goal:4 viewed:2 towards:1 shared:2 fisher:2 change:1 corrected:1 denoising:1 called:2 hospital:1 experimental:1 shannon:1 exception:1 scan:1 collins:1 categorize:1 incorporate:1 |
1,855 | 2,685 | The Laplacian PDF Distance: A Cost
Function for Clustering in a Kernel
Feature Space
Robert Jenssen1?, Deniz Erdogmus2 , Jose Principe2 , Torbj?rn Eltoft1
1
2
Department of Physics, University of Troms?, Norway
Computational NeuroEngineering Laboratory, University of Florida, USA
Abstract
A new distance measure between probability density functions
(pdfs) is introduced, which we refer to as the Laplacian pdf distance. The Laplacian pdf distance exhibits a remarkable connection to Mercer kernel based learning theory via the Parzen window
technique for density estimation. In a kernel feature space defined
by the eigenspectrum of the Laplacian data matrix, this pdf distance is shown to measure the cosine of the angle between cluster
mean vectors. The Laplacian data matrix, and hence its eigenspectrum, can be obtained automatically based on the data at hand,
by optimal Parzen window selection. We show that the Laplacian
pdf distance has an interesting interpretation as a risk function
connected to the probability of error.
1
Introduction
In recent years, spectral clustering methods, i.e. data partitioning based on the
eigenspectrum of kernel matrices, have received a lot of attention [1, 2]. Some
unresolved questions associated with these methods are for example that it is not
always clear which cost function that is being optimized and that is not clear how
to construct a proper kernel matrix.
In this paper, we introduce a well-defined cost function for spectral clustering. This
cost function is derived from a new information theoretic distance measure between
cluster pdfs, named the Laplacian pdf distance. The information theoretic/spectral
duality is established via the Parzen window methodology for density estimation.
The resulting spectral clustering cost function measures the cosine of the angle
between cluster mean vectors in a Mercer kernel feature space, where the feature
space is determined by the eigenspectrum of the Laplacian matrix. A principled
approach to spectral clustering would be to optimize this cost function in the feature
space by assigning cluster memberships. Because of space limitations, we leave it
to a future paper to present an actual clustering algorithm optimizing this cost
function, and focus in this paper on the theoretical properties of the new measure.
?
Corresponding author. Phone: (+47) 776 46493. Email: [email protected]
An important by-product of the theory presented is that a method for learning the
Mercer kernel matrix via optimal Parzen windowing is provided. This means that
the Laplacian matrix, its eigenspectrum and hence the feature space mapping can
be determined automatically. We illustrate this property by an example.
We also show that the Laplacian pdf distance has an interesting relationship to the
probability of error.
In section 2, we briefly review kernel feature space theory. In section 3, we utilize
the Parzen window technique for function approximation, in order to introduce the
new Laplacian pdf distance and discuss some properties in sections 4 and 5. Section
6 concludes the paper.
2
Kernel Feature Spaces
Mercer kernel-based learning algorithms [3] make use of the following idea: via a
nonlinear mapping
? : Rd ? F, x ? ?(x)
(1)
the data x1 , . . . , xN ? Rd is mapped into a potentially much higher dimensional
feature space F. For a given learning problem one now considers the same algorithm
in F instead of in Rd , that is, one works with ?(x1 ), . . . , ?(xN ) ? F.
Consider a symmetric kernel function k(x, y). If k : C ? C ? R is a continuous
kernel of a positive integral operator in a Hilbert space L2 (C) on a compact set
C ? Rd , i.e.
Z
?? ? L2 (C) :
C
k(x, y)?(x)?(y)dxdy ? 0,
(2)
then there exists a space F and a mapping ? : Rd ? F, such that by Mercer?s
theorem [4]
NF
X
k(x, y) = h?(x), ?(y)i =
?i ?i (x)?i (y),
(3)
i=1
where h?, ?i denotes an inner product, the ?i ?s are the orthonormal eigenfunctions
of the kernel and NF ? ? [3]. In this case
p
p
?(x) = [ ?1 ?1 (x), ?2 ?2 (x), . . . ]T ,
(4)
can potentially be realized.
In some cases, it may be desirable to realize this mapping. This issue has been
addressed in [5]. Define the (N ? N ) Gram matrix, K, also called the affinity, or
kernel matrix, with elements Kij = k(xi , xj ), i, j = 1, . . . , N . This matrix can be
diagonalized as ET KE = ?, where the columns of E contains the eigenvectors of K
?1 , . . . , ?
?N , ?
?1 ?
and ? is a diagonal matrix containing the non-negative eigenvalues ?
?
? ? ? ? ?N . In [5], it was shown that the eigenfunctions and eigenvalues of (4) can
?
??
be approximated as ?j (xi ) ? N eji , ?j ? Nj , where eji denotes the ith element
of the jth eigenvector. Hence, the mapping (4), can be approximated as
q
q
?
? N e N i ]T .
?(xi ) ? [ ?1 e1i , . . . , ?
(5)
Thus, the mapping is based on the eigenspectrum of K. The feature space data set
may be represented in matrix form as ?N ?N = [?(x1 ), . . . , ?(xN )]. Hence, ? =
1
? 2 ET . It may be desirable to truncate the mapping (5) to C-dimensions. Thus,
? It is well-known that K
? =?
?T?
? is
only the C first rows of ? are kept, yielding ?.
the best rank-C approximation to K wrt. the Frobenius norm [6].
The most widely used Mercer kernel is the radial-basis-function (RBF)
||x ? y||2
k(x, y) = exp ?
.
2? 2
3
(6)
Function Approximation using Parzen Windowing
Parzen windowing is a kernel-based density estimation method, where the resulting
density estimate is continuous and differentiable provided that the selected kernel
is continuous and differentiable [7]. Given a set of iid samples {x1 , . . . , xN } drawn
from the true density f (x), the Parzen window estimate for this distribution is [7]
N
1 X
?
f(x)
=
W?2 (x, xi ),
(7)
N i=1
where W?2 is the Parzen window, or kernel, and ? 2 controls the width of the kernel.
The Parzen window must integrate to one, and is typically chosen to be a pdf itself
with mean xi , such as the Gaussian kernel
||x ? xi ||2
1
?
W?2 (x, xi ) =
,
(8)
d exp
2? 2
(2?? 2 ) 2
which we will assume in the rest of this paper. In the conclusion, we briefly discuss
the use of other kernels.
Consider a function h(x) = v(x)f (x), for some function v(x). We propose to
estimate h(x) by the following generalized Parzen estimator
N
1 X
?
v(xi )W?2 (x, xi ).
(9)
h(x)
=
N i=1
This estimator is asymptotically unbiased, which can be shown as follows
(
) Z
N
1 X
Ef
v(xi )W?2 (x, xi ) = v(z)f (z)W?2 (x, z)dz = [v(x)f (x)] ? W?2 (x),
N i=1
(10)
where Ef (?) denotes expectation with respect to the density f (x). In the limit as
N ? ? and ?(N ) ? 0, we have
lim [v(x)f (x)] ? W?2 (x) = v(x)f (x).
(11)
N ??
?(N )?0
Of course, if v(x) = 1 ?x, then (9) is nothing but the traditional Parzen estimator
of h(x) = f (x). The estimator (9) is also asymptotically consistent provided that
the kernel width ?(N ) is annealed at a sufficiently slow rate. The proof will be
presented in another paper.
Many approaches have been proposed in order to optimally determine the size of
the Parzen window, given a finite sample data set. A simple selection rule was
proposed by Silverman [8], using the mean integrated square error (MISE) between
the estimated and the actual pdf as the optimality metric:
1
?opt = ?X 4N ?1 (2d + 1)?1 d+4 ,
(12)
P
2
?1
where d is the dimensionality of the data and ?X = d
i ?Xii , where ?Xii are the
diagonal elements of the sample covariance matrix. More advanced approximations
to the MISE solution also exist.
4
The Laplacian PDF Distance
Cost functions for clustering are often based on distance measures between pdfs.
The goal is to assign memberships to the data patterns with respect to a set of
clusters, such that the cost function is optimized.
Assume that a data set consists of two clusters. Associate the probability density
function p(x) with one of the clusters, and the density q(x) with the other cluster.
Let f (x) be the overall probability density function of the dataR set. Now define the
f ?1 weighted inner product between p(x) and q(x) as hp, qif ? p(x)q(x)f ?1 (x)dx.
In such an inner product space, the Cauchy-Schwarz inequality holds, that is,
2
hp, qif ? hp, pif hq, qif . Based on this discussion, an information theoretic distance
measure between the two pdfs can be expressed as
hp, qif
DL = ? log q
? 0.
hp, pif hq, qif
(13)
We refer to this measure as the Laplacian pdf distance, for reasons that we discuss
next. It can be seen that the distance DL is zero if and only if the two densities
are equal. It is non-negative, and increases as the overlap between the two pdfs
decreases. However, it does not obey the triangle inequality, and is thus not a
distance measure in the strict mathematical sense.
We will now show that the Laplacian pdf distance is also a cost function for clustering in a kernel feature space, using the generalized Parzen estimators discussed
in the previous section. Since the logarithm is a monotonic function, we will derive
the expression for the argument of the log in (13). This quantity will for simplicity
be denoted by the letter ?L? in equations.
Assume that we have available the iid data points {xi }, i = 1, . . . , N1 , drawn from
p(x), which is the density of cluster C1 , and the iid {xj }, j = 1, . . . , N2 , drawn from
1
1
q(x), the density of C2 . Let h(x) = f ? 2 (x)p(x) and g(x) = f ? 2 (x)q(x). Hence, we
may write
R
h(x)g(x)dx
q
.
(14)
L= R
R
h2 (x)dx g 2 (x)dx
We estimate h(x) and g(x) by the generalized Parzen kernel estimators, as follows
N1
1 X
1
?
h(x) =
f ? 2 (xi )W?2 (x, xi ),
N1 i=1
N2
1 X
1
g?(x) =
f ? 2 (xj )W?2 (x, xj ).
N2 j=1
(15)
The approach taken, is to substitute these estimators into (14), to obtain
Z
h(x)g(x)dx
N1
N2
1
1
1 X
1 X
f ? 2 (xi )W?2 (x, xi )
f ? 2 (xj )W?2 (x, xj )
N1 i=1
N2 j=1
?
Z
=
Z
NX
1 ,N2
1
? 12
? 12
f (xi )f (xj ) W?2 (x, xi )W?2 (x, xj )dx
N1 N2 i,j=1
=
NX
1 ,N2
1
1
1
f ? 2 (xi )f ? 2 (xj )W2?2 (xi , xj ),
N1 N2 i,j=1
(16)
where in the last step, the convolution theorem for Gaussians has been employed.
Similarly, we have
Z
h2 (x)dx ?
N1 ,N1
1
1 X ?1
f 2 (xi )f ? 2 (xi0 )W2?2 (xi , xi0 ),
2
N1 0
Z
g 2 (x)dx ?
N2 ,N2
1
1 X ?1
f 2 (xj )f ? 2 (xj 0 )W2?2 (xj , xj 0 ).
N22 0
(17)
i,i =1
(18)
j,j =1
Now we define the matrix Kf , such that
1
1
Kfij = Kf (xi , xj ) = f ? 2 (xi )f ? 2 (xj )K(xi , xj ),
(19)
where K(xi , xj ) = W2?2 (xi , xj ) for i, j = 1, . . . , N and N = N1 + N2 . As a
consequence, (14) can be re-written as follows
PN1 ,N2
i,j=1 Kf (xi , xj )
L = qP
(20)
PN2 ,N2
N1 ,N1
0)
0)
K
(x
,
x
K
(x
,
x
0
0
f
i
i
f
j
j
i,i =1
j,j =1
The key point of this paper, is to note that the matrix K = Kij = K(xi , xj ), i, j =
1, . . . , N , is the data affinity matrix, and that K(xi , xj ) is a Gaussian RBF kernel
function. Hence, it is also a kernel function that satisfies Mercer?s theorem. Since
K(xi , xj ) satisfies Mercer?s theorem, the following by definition holds [4]. For any
set of examples {x1 , . . . , xN } and any set of real numbers ?1 , . . . , ?N
N X
N
X
i=1 j=1
?i ?j K(xi , xj ) ? 0,
(21)
in analogy to (3). Moreover, this means that
N X
N
X
1
1
?i ?j f ? 2 (xi )f ? 2 (xj )K(xi , xj ) =
i=1 j=1
N X
N
X
i=1 j=1
?i ?j Kf (xi , xj ) ? 0,
(22)
hence Kf (xi , xj ) is also a Mercer kernel.
Now, it is readily observed that the Laplacian pdf distance can be analyzed in terms
of inner products in a Mercer kernel-based Hilbert feature space, since Kf (xi , xj ) =
h?f (xi ), ?f (xj )i. Consequently, (20) can be written as follows
PN1 ,N2
i,j=1 h?f (xi ), ?f (xj )i
L = qP
PN2 ,N2
N1 ,N1
i,i0 =1 h?f (xi ), ?f (xi0 )i
j,j 0 =1 h?f (xj ), ?f (xj 0 )i
D P
E
PN2
N1
1
1
i=1 ?f (xi ), N2
j=1 ?f (xj )
N1
= rD
ED P
E
N2
1 P N1
1 P N1
1
1 P N2
i=1 ?f (xi ), N1
i0 =1 ?f (xi0 )
j=1 ?f (xj ), N2
j 0 =1 ?f (xj 0 )
N1
N2
m1f , m2f
=
= cos 6 (m1f , m2f ),
(23)
||m1f ||||m2f ||
P Ni
where mif = N1i l=1
?f (xl ), i = 1, 2, that is, the sample mean of the ith cluster
in feature space.
This is a very interesting result. We started out with a distance measure between
densities in the input space. By utilizing the Parzen window method, this distance
measure turned out to have an equivalent expression as a measure of the distance
between two clusters of data points in a Mercer kernel feature space. In the feature
space, the distance that is measured is the cosine of the angle between the cluster
mean vectors.
The actual mapping of a data point to the kernel feature space is given by the
eigendecomposition of Kf , via (5). Let us examine this mapping in more detail.
1
Note that f 2 (xi ) can be estimated from the data by the traditional Parzen pdf
estimator as follows
v
u
N
u1 X
p
1
f 2 (xi ) = t
W?f2 (xi , xl ) = di .
(24)
N
l=1
Define the matrix D = diag(d1 , . . . , dN ). Then Kf can be expressed as
1
1
Kf = D? 2 KD? 2 .
(25)
Quite interestingly, for ?f2 = 2? 2 , this is in fact the Laplacian data matrix.
1
The above discussion explicitly connects the Parzen kernel and the Mercer kernel.
Moreover, automatic procedures exist in the density estimation literature to optimally determine the Parzen kernel given a data set. Thus, the Mercer kernel is
also determined by the same procedure. Therefore, the mapping by the Laplacian
matrix to the kernel feature space can also be determined automatically. We regard
this as a significant result in the kernel based learning theory.
As an example, consider Fig. 1 (a) which shows a data set consisting of a ring
with a dense cluster in the middle. The MISE kernel size is ?opt = 0.16, and
the Parzen pdf estimate is shown in Fig. 1 (b). The data mapping given by the
corresponding Laplacian matrix is shown in Fig. 1 (c) (truncated to two dimensions
for visualization purposes). It can be seen that the data is distributed along two lines
radially from the origin, indicating that clustering based on the angular measure
we have derived makes sense.
The above analysis can easily be extended to any number of pdfs/clusters. In the
C-cluster case, we define the Laplacian pdf distance as
L=
C?1
XX
i=1 j6=i
C
q
hpi , pj if
.
(26)
hpi , pi if hpj , pj if
In the kernel feature space, (26), corresponds to all cluster mean vectors being
pairwise as orthogonal to each other as possible, for all possible unique pairs.
4.1
Connection to the Ng et al. [2] algorithm
Recently, Ng et al. [2] proposed to map the input data to a feature space determined
by the eigenvectors corresponding to the C largest eigenvalues of the Laplacian matrix. In that space, the data was normalized to unit norm and clustered by the
C-means algorithm. We have shown that the Laplacian pdf distance provides a
1
It is a bit imprecise to refer to Kf as the Laplacian matrix, as readers familiar with
spectral graph theory may recognize, since the definition of the Laplacian matrix is L =
I ? Kf . However, replacing Kf by L does not change the eigenvectors, it only changes the
eigenvalues from ?i to 1 ? ?i .
0
0
(a) Data set
(b) Parzen pdf estimate
(c) Feature space data
Figure 1: The kernel size is automatically determined (MISE), yielding the Parzen
estimate (b) with the corresponding feature space mapping (c).
clustering cost function, measuring the cosine of the angle between cluster means,
in a related kernel feature space, which in our case can be determined automatically. A more principled approach to clustering than that taken by Ng et al. is to
optimize (23) in the feature space, instead of using C-means. However, because of
the normalization of the data in the feature space, C-means can be interpreted as
clustering the data based on an angular measure. This may explain some of the
success of the Ng et al. algorithm; it achieves more or less the same goal as clustering based on the Laplacian distance would be expected to do. We will investigate
this claim in our future work. Note that we in our framework may choose to use
only the C largest eigenvalues/eigenvectors in the mapping, as discussed in section
2. Since we incorporate the eigenvalues in the mapping, in contrast to Ng et al.,
the actual mapping will in general be different in the two cases.
5
The Laplacian PDF distance as a risk function
We now give an analysis of the Laplacian pdf distance that may further motivate its
use as a clustering cost function. Consider again the two cluster case. The overall
data distribution can be expressed as f (x) = P1 p(x) + P2 q(x), were Pi , i = 1, 2, are
the priors. Assume that the two clusters are well separated, such that for xi ? C1 ,
f (xi ) ? P1 p(xi ), while for xi ? C2 , f (xi ) ? P2 q(xi ). Let us examine the numerator
R
of (14) in this case. It can be approximated as p(x)q(x)
f (x) dx
Z
Z
Z
Z
p(x)q(x)
1
1
p(x)q(x)
?
dx +
dx ?
q(x)dx +
p(x)dx.
(27)
f (x)
f (x)
P1 C 1
P2 C 2
C1
C2
By performing a similar calculation for the denominator of (14), it can be shown to
be approximately equal to ?P1 P . Hence, the Laplacian pdf distance can be written
1 1
as a risk function, given by
Z
Z
p
1
1
L ? P 1 P2
q(x)dx +
p(x)dx .
(28)
P1 C 1
P2 C 2
Note that if P1 = P2 = 12 , then L = 2Pe , where Pe is the probability of error when
assigning data points to the two clusters, that is
Z
Z
Pe = P 1
q(x)dx + P2
p(x)dx.
(29)
C1
C2
Thus, in this case, minimizing L is equivalent to minimizing Pe . However, in the case
that P1 6= P2 , (28) has an even more interesting interpretation. In that situation,
it can be seen that the two integrals in the expressions (28)Rand (29) are weighted
exactly oppositely. For example, if P1 is close to one, L ? C2 p(x)dx, while Pe ?
R
q(x)dx. Thus, the Laplacian pdf distance emphasizes to cluster the most unC1
likely data points correctly. In many real world applications, this property may be
crucial. For example, in medical applications, the most important points to classify
correctly are often the least probable, such as detecting some rare disease in a group
of patients.
6
Conclusions
We have introduced a new pdf distance measure that we refer to as the Laplacian
pdf distance, and we have shown that it is in fact a clustering cost function in a
kernel feature space determined by the eigenspectrum of the Laplacian data matrix.
In our exposition, the Mercer kernel and the Parzen kernel is equivalent, making
it possible to determine the Mercer kernel based on automatic selection procedures
for the Parzen kernel. Hence, the Laplacian data matrix and its eigenspectrum can
be determined automatically too. We have shown that the new pdf distance has an
interesting property as a risk function.
The results we have derived can only be obtained analytically using Gaussian kernels. The same results may be obtained using other Mercer kernels, but it requires
an additional approximation wrt. the expectation operator. This discussion is left
for future work.
Acknowledgments.
0300340.
This work was partially supported by NSF grant ECS-
References
[1] Y. Weiss, ?Segmentation Using Eigenvectors: A Unifying View,? in International Conference on Computer Vision, 1999, pp. 975?982.
[2] A. Y. Ng, M. Jordan, and Y. Weiss, ?On Spectral Clustering: Analysis and an
Algorithm,? in Advances in Neural Information Processing Systems, 14, 2001,
vol. 2, pp. 849?856.
[3] K. R. M?
uller, S. Mika, G. R?
atsch, K. Tsuda, and B. Sch?
olkopf, ?An Introduction
to Kernel-Based Learning Algorithms,? IEEE Transactions on Neural Networks,
vol. 12, no. 2, pp. 181?201, 2001.
[4] J. Mercer, ?Functions of Positive and Negative Type and their Connection with
the Theory of Integral Equations,? Philos. Trans. Roy. Soc. London, vol. A, pp.
415?446, 1909.
[5] C. Williams and M. Seeger, ?Using the Nystr?
om Method to Speed Up Kernel
Machines,? in Advances in Neural Information Processing Systems 13, Vancouver, Canada, USA, 2001, pp. 682?688.
[6] M. Brand and K. Huang, ?A Unifying Theorem for Spectral Embedding and
Clustering,? in Ninth Int?l Workshop on Artificial Intelligence and Statistics,
Key West, Florida, USA, 2003.
[7] E. Parzen, ?On the Estimation of a Probability Density Function and the
Mode,? Ann. Math. Stat., vol. 32, pp. 1065?1076, 1962.
[8] B. W. Silverman, Density Estimation for Statistics and Data Analysis, Chapman and Hall, London, 1986.
| 2685 |@word middle:1 briefly:2 norm:2 covariance:1 nystr:1 contains:1 interestingly:1 diagonalized:1 assigning:2 dx:19 must:1 written:3 readily:1 realize:1 deniz:1 intelligence:1 selected:1 ith:2 detecting:1 provides:1 math:1 mathematical:1 dn:1 c2:5 along:1 consists:1 troms:1 n22:1 introduce:2 pairwise:1 torbj:1 expected:1 p1:8 examine:2 automatically:6 actual:4 window:9 provided:3 xx:1 moreover:2 interpreted:1 eigenvector:1 nj:1 nf:2 exactly:1 partitioning:1 control:1 unit:1 medical:1 grant:1 positive:2 limit:1 consequence:1 datar:1 approximately:1 mika:1 co:1 kfij:1 unique:1 acknowledgment:1 silverman:2 procedure:3 imprecise:1 radial:1 close:1 selection:3 operator:2 risk:4 optimize:2 equivalent:3 map:1 dz:1 pn2:3 annealed:1 williams:1 attention:1 ke:1 simplicity:1 estimator:8 rule:1 utilizing:1 orthonormal:1 embedding:1 origin:1 associate:1 element:3 roy:1 approximated:3 observed:1 connected:1 decrease:1 principled:2 disease:1 motivate:1 f2:2 basis:1 triangle:1 easily:1 represented:1 separated:1 london:2 artificial:1 quite:1 widely:1 statistic:2 itself:1 eigenvalue:6 differentiable:2 propose:1 product:5 unresolved:1 turned:1 frobenius:1 olkopf:1 cluster:21 leave:1 ring:1 illustrate:1 derive:1 stat:1 measured:1 received:1 p2:8 soc:1 assign:1 clustered:1 opt:2 probable:1 neuroengineering:1 hold:2 sufficiently:1 hall:1 exp:2 mapping:15 claim:1 achieves:1 purpose:1 estimation:6 schwarz:1 largest:2 weighted:2 uller:1 always:1 gaussian:3 derived:3 focus:1 pdfs:6 pif:2 rank:1 contrast:1 seeger:1 sense:2 membership:2 i0:2 typically:1 integrated:1 issue:1 overall:2 denoted:1 equal:2 construct:1 ng:6 chapman:1 future:3 recognize:1 familiar:1 connects:1 consisting:1 n1:21 investigate:1 analyzed:1 hpi:2 yielding:2 integral:3 orthogonal:1 logarithm:1 re:1 tsuda:1 theoretical:1 kij:2 column:1 classify:1 measuring:1 cost:13 rare:1 too:1 optimally:2 density:17 international:1 physic:1 parzen:25 again:1 containing:1 choose:1 huang:1 int:1 explicitly:1 view:1 lot:1 om:1 square:1 ni:1 m2f:3 emphasizes:1 iid:3 j6:1 explain:1 phys:1 ed:1 email:1 definition:2 pp:6 associated:1 proof:1 di:1 mise:4 radially:1 lim:1 dimensionality:1 hilbert:2 segmentation:1 norway:1 higher:1 oppositely:1 methodology:1 wei:2 rand:1 angular:2 hand:1 replacing:1 nonlinear:1 mode:1 usa:3 normalized:1 true:1 unbiased:1 hence:9 analytically:1 symmetric:1 laboratory:1 hpj:1 width:2 numerator:1 cosine:4 generalized:3 pdf:26 theoretic:3 ef:2 recently:1 mif:1 qp:2 discussed:2 interpretation:2 xi0:4 refer:4 significant:1 pn1:2 rd:6 automatic:2 philos:1 hp:5 similarly:1 recent:1 optimizing:1 phone:1 inequality:2 success:1 seen:3 dxdy:1 additional:1 employed:1 determine:3 windowing:3 desirable:2 eji:2 calculation:1 laplacian:31 denominator:1 patient:1 expectation:2 metric:1 vision:1 kernel:49 normalization:1 c1:4 addressed:1 crucial:1 sch:1 w2:4 rest:1 eigenfunctions:2 strict:1 n1i:1 jordan:1 xj:36 inner:4 idea:1 expression:3 clear:2 eigenvectors:5 exist:2 nsf:1 estimated:2 correctly:2 xii:2 write:1 vol:4 group:1 key:2 drawn:3 pj:2 utilize:1 kept:1 asymptotically:2 graph:1 year:1 jose:1 angle:4 letter:1 named:1 reader:1 bit:1 u1:1 speed:1 argument:1 optimality:1 performing:1 department:1 truncate:1 kd:1 making:1 taken:2 equation:2 visualization:1 discus:3 wrt:2 available:1 gaussians:1 obey:1 spectral:8 florida:2 substitute:1 denotes:3 clustering:17 unifying:2 question:1 realized:1 quantity:1 diagonal:2 traditional:2 exhibit:1 affinity:2 hq:2 distance:32 mapped:1 nx:2 considers:1 eigenspectrum:8 cauchy:1 reason:1 relationship:1 minimizing:2 qif:5 robert:1 potentially:2 negative:3 proper:1 convolution:1 finite:1 truncated:1 situation:1 extended:1 rn:1 ninth:1 canada:1 introduced:2 pair:1 connection:3 optimized:2 established:1 trans:1 pattern:1 overlap:1 advanced:1 started:1 concludes:1 review:1 literature:1 l2:2 prior:1 kf:12 vancouver:1 interesting:5 limitation:1 analogy:1 remarkable:1 h2:2 integrate:1 eigendecomposition:1 consistent:1 mercer:17 pi:2 row:1 course:1 supported:1 last:1 jth:1 distributed:1 regard:1 dimension:2 xn:5 uit:1 gram:1 world:1 author:1 ec:1 transaction:1 compact:1 e1i:1 xi:51 continuous:3 diag:1 dense:1 n2:21 nothing:1 x1:5 fig:3 west:1 slow:1 xl:2 pe:5 theorem:5 dl:2 exists:1 workshop:1 likely:1 expressed:3 partially:1 monotonic:1 corresponds:1 satisfies:2 goal:2 consequently:1 rbf:2 exposition:1 ann:1 change:2 determined:9 called:1 duality:1 brand:1 atsch:1 indicating:1 incorporate:1 d1:1 |
1,856 | 2,686 | Efficient Kernel Discriminant Analysis via QR
Decomposition
Tao Xiong
Department of ECE
University of Minnesota
[email protected]
Jieping Ye
Department of CSE
University of Minnesota
[email protected]
Vladimir Cherkassky
Department of ECE
University of Minnesota
[email protected]
Qi Li
Department of CIS
University of Delaware
[email protected]
Ravi Janardan
Department of CSE
University of Minnesota
[email protected]
Abstract
Linear Discriminant Analysis (LDA) is a well-known method for feature extraction and dimension reduction. It has been used widely in
many applications such as face recognition. Recently, a novel LDA algorithm based on QR Decomposition, namely LDA/QR, has been proposed,
which is competitive in terms of classification accuracy with other LDA
algorithms, but it has much lower costs in time and space. However,
LDA/QR is based on linear projection, which may not be suitable for data
with nonlinear structure. This paper first proposes an algorithm called
KDA/QR, which extends the LDA/QR algorithm to deal with nonlinear data by using the kernel operator. Then an efficient approximation of
KDA/QR called AKDA/QR is proposed. Experiments on face image data
show that the classification accuracy of both KDA/QR and AKDA/QR
are competitive with Generalized Discriminant Analysis (GDA), a general kernel discriminant analysis algorithm, while AKDA/QR has much
lower time and space costs.
1
Introduction
Linear Discriminant Analysis [3] is a well?known method for dimension reduction. It has
been used widely in many applications such as face recognition [2]. Classical LDA aims
to find optimal transformation by minimizing the within-class distance and maximizing
the between-class distance simultaneously, thus achieving maximum discrimination. The
optimal transformation can be readily computed by computing the eigen-decomposition on
the scatter matrices.
Although LDA works well for linear problems, it may be less effective when severe nonlinearity is involved. To deal with such a limitation, nonlinear extensions through kernel
functions have been proposed. The main idea of kernel-based methods is to map the input
data to a feature space through a nonlinear mapping, where the inner products in the feature
space can be computed by a kernel function without knowing the nonlinear mapping explicitly [9]. Kernel Principal Component Analysis (KPCA) [10], Kernel Fisher Discriminant
Analysis (KFDA) [7] and Generalized Discriminant Analysis (GDA) [1] are, respectively,
kernel-based nonlinear extensions of the well known PCA, FDA and LDA methods.
To our knowledge, there are few efficient algorithms for general kernel based discriminant
algorithms ? most known algorithms effectively scale as O(n3 ) where n is the sample
size. In [6, 8], S. Mika et al. made a first attempt to speed up KFDA through a greedy
approximation technique. However the algorithm was developed to handle the binary classification problem. For multi-class problem, the authors suggested the one against the rest
scheme by considering all two-class problems.
Recently, an efficient variant of LDA, namely LDA/QR, was proposed in [11, 12]. The
essence of LDA/QR is the utilization of QR-decomposition on a small size matrix. The
time complexity of LDA/QR is linear in the size of the training data, as well as the number
of dimensions of the data. Moreover, experiments in [11, 12] show that the classification
accuracy of LDA/QR is competitive with other LDA algorithms.
In this paper, we first propose an algorithm, namely KDA/QR1 , which is a nonlinear extension of LDA/QR. Since KDA/QR involves the whole kernel matrix, which is not scalable
for large datasets, we also propose an approximation of KDA/QR, namely AKDA/QR. A
distinct property of AKDA/QR is that it scales as O(ndc), where n is the size of the data,
d is the dimension of the data, and c is the number of classes.
We apply the proposed algorithms on face image datasets and compare them with LDA/QR,
and Generalized Discriminant Analysis (GDA) [1], a general method for kernel discriminant analysis. Experiments show that: (1) AKDA/QR is competitive with KDA/QR and
GDA in classification; (2) both KDA/QR and AKDA/QR outperform LDA/QR in classification; and (3) AKDA/QR has much lower costs in time and space than GDA.
2
LDA/QR
In this section, we give a brief review of the LDA/QR algorithm [11, 12]. This algorithm
has two stages. The first stage maximizes the separation between different classes via QR
Decomposition [4]. The second stage addresses the issue of minimizing the within-class
distance, while maintaining low time/space complexity.
Let A ? IRd?n be the data matrix, where each column ai is a vector in d-dimensional space.
Assume A is partitioned into c classes {?i }ci=1 , and the size of the ith class |?i | = ni .
Define between-class, within-class, and total scatter matrices Sb , Sw , and St respectively,
?
as follows?[3]: Sb = Hb Hbt , Sw = Hw Hwt , and St = Ht Htt , where Hb = [ N1 (m1 ?
m), ? ? ? , Nc (mc ? m)] ? Rd?c , Hw = A ? [m1 et1 , ? ? ? , mc etc ] ? Rd?n , and Ht =
A ? met ? Rd?n , ei = (1, ? ? ? , 1)t ? Rni ?1 , e = (1, ? ? ? , 1)t ? Rn?1 , mi is the mean of
the ith class, and m is the global mean. It is easy to check that St = Sb + Sw .
The first stage of LDA/QR aims to solve the following optimization problem,
G = arg max
trace(Gt Sb G).
t
G G=I
(1)
Note that this optimization only addresses the issue of maximizing the between-class distance. The solution can be obtained by solving the eigenvalue problem on Sb . The solution
can also be obtained through QR Decomposition on the centroid matrix C [12], where
C = [m1 , m2 , ? ? ? , mc ] consists of the c centroids. More specifically, let C = QR be
the QR Decomposition of C, where Q ? IRn?c has orthonormal columns and R ? IRc?c
1
KDA/QR stands for Kernel Discriminant Analysis via QR-decomposition
Algorithm 1: LDA/QR
/* Stage I: */
1. Construct centroid matrix C;
2. Compute QR Decomposition of C as C = QR, where Q ? IRd?c , R ? IRc?c ;
/* Stage II: */
3. Y ? Hbt Q;
4. Z ? Htt Q;
5. B ? Y t Y ; /*Reduced between-class scatter matrix*/
6. T ? Z t Z; /*Reduced total scatter matrix*/
7. Compute the c eigenvectors ?i of (T + ?Ic )?1 B with decreasing eigenvalues;
8. G ? QV , where V = [?1 , ? ? ? , ?c ].
is upper triangular. Then G = QV , for any orthogonal matrix V , solves the optimization problem in Eq. (1). Note that the choice of orthogonal matrix V is arbitrary, since
trace(Gt Sb G) = trace(V t Gt Sb GV ), for any orthogonal matrix V .
The second stage of LDA/QR refines the first stage by addressing the issue of minimizing
the within-class distance. It incorporates the within-class scatter information by applying a
relaxation scheme on V (relaxing V from an orthogonal matrix to an arbitrary matrix). In
the second stage of LDA/QR, we look for a transformation matrix G such that G = QV , for
some V . Note that V is not required to be orthogonal. The original problem of computing G
is equivalent to computing V . Since Gt Sb G = V t (Qt Sb Q)V , Gt Sw G = V t (Qt Sw Q)V ,
and Gt St G = V t (Qt St Q)V , the original problem of finding optimal G is equivalent to
finding V , with B = Qt Sb Q, W = Qt Sw Q, and T = Qt St Q as the ?reduced? betweenclass, within-class and total scatter matrices, respectively. Note that B has much smaller
size than the original scatter matrix Sb (similarly for W and T ).
The optimal V can be computed efficiently using many existing LDA-based methods, since
we are dealing with matrices B, W , and T of size c by c. We can compute the optimal V
by simply applying regularized LDA; that is, we compute V , by solving a small eigenvalue
problem on (W + ?Ic )?1 B or (T + ?Ic )?1 B (note T = B + W ), for some positive
constant ? [3]. The pseudo-code for this algorithm is given in Algorithm 1. We use the
total scatter instead of the within-class scatter in Lines 4, 6, and 7, mainly for convenience
of presentation of the kernel methods in Section 3 and Section 4.
3
Kernel discriminant analysis via QR-decomposition (KDA/QR)
In this section, the KDA/QR algorithm, a nonlinear extension of LDA/QR through kernel
functions, is presented. Let ? be a mapping to the feature space and ?(A) be the data
matrix in the feature space. Then, the centroid matrix C ? in the feature space is
?
1
1
?
?
?(ai ), ? ? ? ,
?(ai ) .
(2)
C = m1 , ? ? ? , mc =
n1
nc
i??1
i??c
The global centroid in the feature space can be computed as m? = n1 i ni m?
i . To maximize between-class distance in the feature space, as discussed in Section 2, we perform QR
decomposition on C ? , i.e., C ? = Q? R? . A key observation is that R? can be computed
as (C ? )t C ? = (R? )t R? by applying the Cholesky decomposition on (C ? )t C ? [4].
Note that C ? = A? M , where A? = ?(A) = [?(a1 ) . . . ?(an )], and the ith column
of M is (0, ? ? ? , 0, n1i , ? ? ? , n1i , 0, ? ? ? , 0)t . Let K be the kernel matrix with K(i, j) =
?(ai ), ?(aj ). Then
(C ? )t C ? = M t KM.
(3)
Algorithm 2: KDA/QR
/* Stage I: */
1. Construct kernel matrix K;
2. Compute (C ? )t C ? = M t (KM ) as in Eq. (3);
3. Compute R? from the Cholesky Decomposition of (C ? )t C ? ;
/* Stage II: */
4. Y ? ? N t M t KM (R? )?1 ;
5. Z ? ? E t KM (R? )?1 ;
6. B ? ? (Y ? )t Y ? ;
7. T ? ? (Z ? )t Z ? ;
?
?1 ?
B , with decreasing eigenvalues;
8. Compute the c eigenvectors ??
i of (T + ?Ic )
?
?
,
?
,
?
?
?
,
?
];
9. V ? ? [??
c
1
2
10. G? ? C ? (R? )?1 V ? ;
With the computed R? , Q? = C ? (R? )?1 . The matrices Y ? , Z ? , B ? , and W ? in the
feature space (corresponding to the second stage in LDA/QR) can be computed as follows.
In the feature space, we
have Hb? = C ? N , where the ith column of N is
?
?
n
((0, ? ? ? , ni , ? ? ? 0)t ? n i (n1 , ? ? ? , nc )t . It follows that Y ? = (Hb? )t Q? =
N t (C ? )t C ? (R? )?1 = N t M t KM (R? )?1 . Similarly, Ht? = A? E and Z ? =
(Ht? )t Q? = E t (A? )t C ? (R? )?1 = E t (A? )t A? M (R? )?1 = E t KM (R? )?1 , where
E = I ? n1 eet .
Since Sb? = Hb? (Hb? )t and St? = Ht? (Ht? )t , we have
B?
T?
= (Q? )t Sb? Q? = (Q? )t Hb? (Hb? )t Q? = (Y ? )t Y ? ,
= (Q? )t St? Q? = (Q? )t Ht? (Ht? )t Q? = (Z ? )t Z ? .
c
?
?1 ?
We proceed by computing the c eigenvectors {??
B . Define V ? =
i }i=1 of (T + ?Ic )
?
?
?
[?1 , ?2 , ? ? ? , ?c ]. The final transformation matrix can be computed as
G? = Q? V ? = C ? (R? )?1 V ? .
(4)
=
For a given data point z, its projection by G? is (G? )t ?(z)
(V ? )t ((R? )?1 )t (C ? )t ?(z) = (V ? )t ((R? )?1 )t M t Ktz , where Ktz ? IRn and
Ktz (i) = ?(ai ), ?(z).
The pseudo-code for the KDA/QR algorithm is given in Algorithm 2.
3.1
Complexity analysis of KDA/QR
The cost to formulate the kernel matrix in Line 1 is O(n2 d). The computation of (C ? )t C ?
in Line 2 takes O(n2 ), taking advantage of the sparse structure of M . The Cholesky decomposition in Line 3 takes O(c3 ) [4]. Lines 4 takes O(c3 ), as M t KM is already computed in
Line 2. In Line 5, the
computation of Z ?
= E t KM (R? )?1 = (I ? n1 eet )KM (R? )?1 =
KM (R? )?1 ? n1 e (et KM )(R? )?1 in the given order takes O(nc2 ), assuming KM
is kept in Line 2. Lines 6, 7, and 8 take O(c3 ), O(nc2 ) and O(c3 ), respectively. Hence,
the total complexity of the kernel LDA/QR algorithm is O(n2 d). Omitting the cost for
evaluating the kernel matrix K, which is required in all kernel-based algorithms, the total
cost is O(n2 ). Note that all other general discriminant analysis algorithms scale as O(n3 ).
4
Approximate KDA/QR (AKDA/QR)
In this section, we present the AKDA/QR algorithm, which is an efficient approximation
of the KDA/QR algorithm from the last section. Note that the bottleneck of KDA/QR is the
explicit formation of the large kernel matrix K for the computation of (C ? )t C ? in Line
2 of Algorithm 2. The AKDA/QR algorithm presented in this section avoids the explicit
construction of K, thus reducing the computational cost significantly.
? t ?
?
The key to AKDA/QR is the
=
efficient computation of (C ) C , ?where C
1
?
?
?
[m1 , ? ? ? , mc ] and mj = nj i??j ?(ai ). AKDA/QR aims to find xj in the original
?
space such that ?(x?j ) approximates m?
j . Mathematically, the optimal xj can be computed
by solving the following optimization problem:
1
min ?(xj ) ?
?(ai )2 for j = 1, ? ? ? , c.
(5)
d
nj
xj ?R
i??j
To proceed, we only consider Gaussian kernels for AKDA/QR, as they are the most widely
used ones in the literature [9]. Furthermore, the optimization problem in (5) can be simplified by focusing on the Gaussian kernels, as shown in the following lemma.
Lemma 4.1. Consider Gaussian kernel function exp(?x ? y2 /?), where ? is the bandwidth parameter. The optimization problem in (5) is convex if
2
(xj ? ai ) ? 1
for each j = 1, ? ? ? , c and for all i ? ?j ,
(6)
?
Proof. It is easy to check that, for the Gaussian kernel, the optimization problem in (5)
reduces to:
min f (xj )
xj ?Rd
for j = 1, ? ? ? , c,
(7)
where f (x) = i??j fi (x) and fi (x) = ?exp(?x ? ai 2 /?). The Hessian matrix of
fi (x) is H(fi ) = ?2 exp(?x ? ai 2 /?)(I ? ?2 (x ? ai )(x ? ai )t ). It is easy to show that
if ?2 (x ? ai ) ? 1, for all i ? ?j , then H(fi ) is positive semi-definite, that is, fi (x) is
convex. Thus, f (x), the sum of convex functions is also convex.
For applications involving high-dimensional data, such as face recognition, ? is usually
large (typically ranging from thousands to hundreds of thousands [13]), and the condition
in Lemma 4.1 holds if we restrict our search space to the convex hull of each class in the
original space. Therefore, the global minimum of the optimization problem in (7) can be
found very efficiently using Newton?s or gradient decent methods. A key observation is
that for relatively large ?, the centroid of each class in the original space will map very
close to the centroid in the feature space [9], which can serve as the approximatesolution
of the optimization problem in (7). Experiments show that choosing x?j = n1j i??j ai
produces results close to the one by solving the optimization problem in (7). We thus use
it in all the following experiments.
With the computed x?j , for j = 1, . . . , c, the centroid matrix C ? can be approximated by
C ? ? [?(x?1 ) . . . ?(x?c )] (? C? ? )
(8)
?
(C? ? )t C? ? = K,
(9)
and
Algorithm 3: AKDA/QR
/* Stage I: */
1. Compute x?j = n1j i??j ai , for j = 1, ? ? ? , c;
? as in Eq. (9);
2. Construct kernel matrix K
? ? from the Cholesky Decomposition of K;
?
3. Compute R
/* Stage II: */
? R
? ? )?1 ;
4. Y? ? ? N t K(
? ? )?1 ;
? tc (R
5. Z? ? ? E t K
?
? t ??
?
?
6. B ? (Y ) Y ;
7. T?? ? (Z? ? )t Z? ? ;
?1 ? ?
??
8. Compute the c eigenvectors ???
B , with decreasing eigenvalues;
i of (T + ?Ic )
?
?
?
?
9. V? ? [??1 , ??2 , ? ? ? , ??c ];
? ? ? C? ? (R
? ? )?1 V? ? ;
10. G
time
space
PCA
O(n2 d)
O(nd)
LDA/QR
O(ndc)
O(nc)
GDA
O(n2 d + n3 )
O(n2 )
KDA/QR
O(n2 d)
O(n2 )
AKDA/QR
O(ndc)
O(nc)
Table 1: Comparison of time & space complexities of several dimension reduction algorithms: n is the size of the data, d is the dimension, and c is the number of classes.
? j) = ?(x? ), ?(x? ) and K
? ? Rc?c . The Cholesky decomposition of K
? will
where K(i,
i
j
?
?
t
?
? .
? = (R
? )R
? by K
give us R
? ? )?1 ,
? ? = C? ? N , and Y? ? = N t K(
? R
? ? )?1 . Similarly, Z? ? = E t K
? tc (R
It follows that H
b
?
? tc (i, j) = ?(ai ), ?(x ).
where N and E are defined as in Section 3, and K
j
The following steps will be the same as the KDA/QR algorithm. The pseudo-code for
AKDA/QR is given in Algorithm 3.
4.1
Complexity analysis of AKDA/QR
? in Line 2 takes O(c2 d). The
It takes O(dn) in Line 1. The construction of the matrix K
3
Cholesky Decomposition in Line 3 takes O(c ) [4]. Lines 4 and 5 take O(c3 ) and O(ndc)
respectively. It then takes O(c3 ) and O(nc2 ) for matrix multiplications in Lines 6 and 7,
respectively. Line 8 computes the eigen-decomposition of a c by c matrix, hence takes
O(c3 ) [4]. Thus, the most expensive step in Algorithm 3 is Line 5, which takes O(ndc).
Table 1 lists the time and space complexities of several dimension reduction algorithms. It
is clear from the table that AKDA/QR is more efficient than other kernel based methods.
5
Experimental results
In this section, we evaluate both the KDA/QR and AKDA/QR algorithms. The performance is measured by classification accuracy. Note that both KDA/QR and AKDA/QR
have two parameters: ? for the kernel function and ? for the regularization. Experiments show that choosing ? = 100000 and ? = 0.15 for KDA/QR, and ? = 100000
and ? = 0.10 for AKDA/QR produce good overall results. We thus use these values
in all the experiments. 1-Nearest Neighbor (1-NN) method is used as the classifier. We
randomly select p samples of each person from the dataset for training and the rest for
0.9
0.99
0.8
0.98
0.7
Accuracy
Accuracy
1
0.97
0.96
PCA
LDA/QR
KDA/QR
AKDA/QR
0.95
0.6
PCA
LDA/QR
KDA/QR
AKDA/QR
0.5
0.4
0.94
0.3
0.93
0.2
3
4
5
6
7
Number of training samples per class
8
3
4
5
6
7
Number of training samples per class
8
Figure 1: Comparison of classification accuracy on PIX (left) and AR (right).
testing. We repeat the experiments 20 times and report the average recognition accuracy
of each method. The MATLAB codes for the KDA/QR and AKDA/QR algorithms may be
accessed at http://www.cs.umn.edu/?jieping/Kernel.
Datasets: We use the following three datasets in our study, which are publicly available:
PIX contains 300 face images of 30 persons. The image size of PIX image is 512 ? 512.
We subsample the images down to a size of 100 ? 100 = 10000; ORL is a well-known
dataset for face recognition. It contains ten different face images of 40 persons, for a total
of 400 images. The image size is 92 ? 112 = 10304; AR is a large face image datasets. We
use a subset of AR. This subset contains 1638 face images of 126 persons. Its image size
is 768 ? 576. We subsample the images down to a size of 60 ? 40 = 2400. Each dataset is
normalized to have zero mean and unit variance.
KDA/QR and AKQA/QR vs. LDA/QR: In this experiment, we compare the performance of AKDA/QR and KDA/QR with that of several other linear dimension reduction
algorithms including PCA, LDA/QR on two face datasets. We use 100 principal components for PCA as it produces good overall results. The results are summarized in Fig. 1,
where the x-axis denotes the number of samples per class in the training set and the y-axis
denotes the classification accuracy. Fig. 1 shows that KDA/QR and AKQA/QR consistently outperform LDA/QR and PCA. The most interesting result lies in the AR dataset,
where AKDA/QR and KDA/QR outperform LDA/QR by a large margin. It is known that
the images in the AR dataset contain pretty large area of occlusion due to sun glasses and
scarves, which makes linear algorithms such as LDA/QR less effective. Another interesting observation is that the approximate AKQA/QR algorithm is competitive with its exact
version KDA/QR in all cases.
KDA/QR and AKQA/QR vs. GDA: In this experiment, we compare the performance of
AKDA/QR and KDA/QR with Generalized Discriminant Analysis (GDA) [1]. The comparison is made on the ORL face dataset, as the result of GDA on ORL is available in
[5]. We also include the results on PCA and LDA/QR. The results are summarized in Table 2. The main observation from this experiment is that both KDA/QR and AKDA/QR are
competitive with GDA, while AKDA/QR is much more efficient than GDA (see Table 1).
Similar to the first experiment, Table 2 shows that KDA/QR and AKDA/QR consistently
outperform the PCA and LDA/QR algorithms in terms of recognition accuracy.
6
Conclusions
In this paper, we first present a general kernel discriminant analysis algorithm, called
KDA/QR. Using Gaussian kernels, we then proposed an approximate algorithm to
p
3
4
5
6
7
8
PCA
0.8611
0.8938
0.9320
0.9512
0.9633
0.9713
LDA/QR
0.8561
0.9083
0.9385
0.9444
0.9692
0.9713
GDA
0.8782
0.9270
0.9535
0.9668
0.9750
0.9938
KDA/QR
0.9132
0.9321
0.9625
0.9737
0.9825
0.9875
AKDA/QR
0.9118
0.9300
0.9615
0.9744
0.9815
0.9875
Table 2: Comparison of classification accuracy on ORL face image dataset. p is the number
of training samples per class. The results on GDA are taken from [5].
KDA/QR, which we call AKDA/QR. Our experimental results show that the accuracy
achieved by the two algorithms is very competitive with GDA, a general kernel discriminant algorithms, while AKDA/QR is much more efficient. In particular, the computational
complexity of AKDA/QR is linear in the number of the data points in the training set as
well as the number of dimensions and the number of classes.
Acknowledgment Research of J. Ye and R. Janardan is sponsored, in part, by the Army High Performance Computing Research Center under the auspices of the Department of the Army, Army
Research Laboratory cooperative agreement number DAAD19-01-2-0014, the content of which does
not necessarily reflect the position or the policy of the government, and no official endorsement
should be inferred.
References
[1] G. Baudat and F. Anouar. Generalized discriminant analysis using a kernel approach. Neural
Computation, 12(10):2385?2404, 2000.
[2] P.N. Belhumeour, J.P. Hespanha, and D.J. Kriegman. Eigenfaces vs. fisherfaces: Recognition
using class specific linear projection. IEEE TPAMI, 19(7):711?720, 1997.
[3] K. Fukunaga. Introduction to Statistical Pattern Classification. Academic Press, San Diego,
California, USA, 1990.
[4] G. H. Golub and C. F. Van Loan. Matrix Computations. The Johns Hopkins University Press,
Baltimore, MD, USA, third edition, 1996.
[5] Q. Liu, R. Huang, H. Lu, and S. Ma. Kernel-based optimized feature vectors selection and
discriminant analysis for face recognition. In ICPR Proceedings, pages 362 ? 365, 2002.
[6] S. Mika, G. R?atsch, and K.-R. M?uller. A mathematical programming approach to the kernel
fisher algorithm. In NIPS Proceedings, pages 591 ? 597, 2001.
[7] S. Mika, G. Ratsch, J. Weston, B. Sch?okopf, and K.-R. M?uller. Fisher discriminant analysis
with kernels. In IEEE Neural Networks for Signal Processing Workshop, pages 41 ? 48, 1999.
[8] S. Mika, A.J. Smola, and B. Sch?olkopf. An improved training algorithm for kernel fisher discriminants. In AISTATS Proceedings, pages 98?104, 2001.
[9] B. Sch?okopf and A. Smola. Learning with Kernels: Support Vector Machines, Regularization,
Optimization and Beyond. MIT Press, 2002.
[10] B. Sch?okopf, A. Smola, and K. M?uller. Nonlinear component analysis as a kernel eigenvalue
problem. Neural Computation, 10(5):1299?1319, 1998.
[11] J. Ye and Q. Li. LDA/QR: An efficient and effective dimension reduction algorithm and its
theoretical foundation. Pattern recognition, pages 851?854, 2004.
[12] J. Ye, Q. Li, H. Xiong, H. Park, R. Janardan, and V. Kumar. IDR/QR: An incremental dimension
reduction algorithm via QR decomposition. In ACM SIGKDD Proceedings, pages 364?373,
2004.
[13] W. Zheng, L. Zhao, and C. Zou. A modified algorithm for generalized discriminant analysis.
Neural Computation, 16(6):1283?1297, 2004.
| 2686 |@word version:1 nd:1 km:12 decomposition:19 reduction:7 liu:1 contains:3 existing:1 scatter:9 readily:1 john:1 refines:1 gv:1 sponsored:1 discrimination:1 v:3 greedy:1 ith:4 cse:2 accessed:1 rc:1 dn:1 c2:1 mathematical:1 consists:1 idr:1 multi:1 decreasing:3 considering:1 moreover:1 maximizes:1 developed:1 finding:2 transformation:4 n1j:2 nj:2 pseudo:3 anouar:1 classifier:1 utilization:1 unit:1 positive:2 mika:4 relaxing:1 acknowledgment:1 testing:1 definite:1 daad19:1 area:1 significantly:1 projection:3 janardan:4 convenience:1 close:2 selection:1 operator:1 baudat:1 applying:3 www:1 equivalent:2 map:2 jieping:3 maximizing:2 center:1 convex:5 formulate:1 m2:1 orthonormal:1 handle:1 construction:2 diego:1 exact:1 programming:1 agreement:1 recognition:9 approximated:1 expensive:1 cooperative:1 thousand:2 sun:1 complexity:8 kriegman:1 solving:4 serve:1 distinct:1 effective:3 formation:1 choosing:2 widely:3 solve:1 triangular:1 final:1 advantage:1 eigenvalue:6 tpami:1 propose:2 product:1 udel:1 olkopf:1 qr:120 produce:3 incremental:1 measured:1 nearest:1 qt:6 eq:3 solves:1 c:3 involves:1 met:1 hull:1 government:1 mathematically:1 extension:4 hold:1 ic:6 exp:3 mapping:3 hwt:1 et1:1 qv:3 uller:3 mit:1 gaussian:5 aim:3 modified:1 consistently:2 check:2 mainly:1 sigkdd:1 centroid:8 glass:1 scarf:1 nn:1 sb:13 typically:1 irn:2 tao:1 arg:1 classification:11 issue:3 overall:2 proposes:1 construct:3 extraction:1 park:1 look:1 report:1 few:1 randomly:1 simultaneously:1 occlusion:1 n1:7 attempt:1 zheng:1 severe:1 golub:1 umn:5 orthogonal:5 theoretical:1 column:4 ar:5 kpca:1 cost:7 addressing:1 subset:2 hundred:1 st:8 person:4 hopkins:1 okopf:3 reflect:1 huang:1 zhao:1 li:3 summarized:2 explicitly:1 competitive:7 nc2:3 ni:3 accuracy:12 publicly:1 variance:1 efficiently:2 mc:5 lu:1 against:1 involved:1 proof:1 mi:1 dataset:7 knowledge:1 focusing:1 improved:1 furthermore:1 stage:14 smola:3 ei:1 nonlinear:9 aj:1 lda:41 usa:2 omitting:1 ye:4 normalized:1 contain:1 hence:2 regularization:2 laboratory:1 deal:2 essence:1 generalized:6 image:15 ranging:1 novel:1 recently:2 fi:6 discriminants:1 discussed:1 m1:5 approximates:1 ai:16 rd:4 similarly:3 nonlinearity:1 minnesota:4 etc:1 gt:6 binary:1 minimum:1 ktz:3 maximize:1 signal:1 ii:3 semi:1 reduces:1 academic:1 a1:1 qi:1 variant:1 scalable:1 involving:1 kernel:41 achieved:1 baltimore:1 ratsch:1 sch:4 rest:2 n1i:2 incorporates:1 call:1 htt:2 easy:3 decent:1 hb:8 xj:7 bandwidth:1 restrict:1 inner:1 idea:1 delaware:1 knowing:1 bottleneck:1 pca:10 ird:2 proceed:2 hessian:1 matlab:1 clear:1 eigenvectors:4 ten:1 reduced:3 http:1 outperform:4 per:4 key:3 achieving:1 ravi:1 ht:8 kept:1 relaxation:1 sum:1 pix:3 extends:1 separation:1 betweenclass:1 endorsement:1 orl:4 n3:3 fda:1 auspex:1 speed:1 min:2 fukunaga:1 kumar:1 relatively:1 department:6 icpr:1 smaller:1 partitioned:1 gda:14 taken:1 kda:37 available:2 apply:1 xiong:2 eigen:2 original:6 denotes:2 include:1 maintaining:1 sw:6 newton:1 classical:1 already:1 md:1 gradient:1 distance:6 discriminant:20 kfda:2 assuming:1 code:4 minimizing:3 vladimir:1 nc:5 trace:3 hespanha:1 policy:1 fisherfaces:1 perform:1 upper:1 observation:4 datasets:6 rn:1 arbitrary:2 inferred:1 namely:4 required:2 c3:7 optimized:1 california:1 nip:1 address:2 beyond:1 suggested:1 usually:1 pattern:2 max:1 including:1 suitable:1 irc:2 regularized:1 scheme:2 brief:1 axis:2 review:1 literature:1 multiplication:1 interesting:2 limitation:1 foundation:1 rni:1 repeat:1 last:1 neighbor:1 eigenfaces:1 face:14 taking:1 sparse:1 van:1 hbt:2 dimension:11 stand:1 evaluating:1 avoids:1 computes:1 author:1 made:2 san:1 simplified:1 approximate:4 eet:2 dealing:1 global:3 search:1 pretty:1 table:7 mj:1 necessarily:1 zou:1 ndc:5 official:1 aistats:1 main:2 whole:1 subsample:2 edition:1 n2:9 fig:2 position:1 explicit:2 lie:1 third:1 hw:2 down:2 specific:1 list:1 workshop:1 effectively:1 ci:3 margin:1 cherkassky:1 tc:3 simply:1 army:3 acm:1 ma:1 weston:1 presentation:1 fisher:4 content:1 loan:1 specifically:1 reducing:1 principal:2 lemma:3 called:3 total:7 ece:4 experimental:2 atsch:1 select:1 cholesky:6 support:1 evaluate:1 |
1,857 | 2,687 | Multi-agent Cooperation
in Diverse Population Games
K. Y. Michael Wong, S. W. Lim and Z. Gao
Hong Kong University of Science and Technology, Hong Kong, China.
{phkywong, swlim, zhuogao}@ust.hk
Abstract
We consider multi-agent systems whose agents compete for resources by
striving to be in the minority group. The agents adapt to the environment
by reinforcement learning of the preferences of the policies they hold.
Diversity of preferences of policies is introduced by adding random biases to the initial cumulative payoffs of their policies. We explain and
provide evidence that agent cooperation becomes increasingly important
when diversity increases. Analyses of these mechanisms yield excellent
agreement with simulations over nine decades of data.
1
Introduction
In the intelligent control of large systems, the multi-agent approach has the advantages of
parallelism, robustness, scalability, and light communication overhead [1]. Since it involves
many interacting adaptive agents, the behavior becomes highly complex. While a standard
analytical approach is to study their steady state behavior described by the Nash equilibria
[2], it is interesting to consider the dynamics of how the steady state is approached. Of
particular interest is the case of heterogeneous agents, which have diversi?ed preferences
in decision making. In such cases, the cooperation of agents becomes very important.
Speci?cally, we consider the dynamics of a version of large population games which models the collective behavior of agents simultaneously and adaptively competing for limited
resources. The game is a variant of the Minority Game, in which the agents strive to
make the minority decision, thereby balancing the load distributed between the majority
and minority choices [3]. Previous work showed that the system behavior depends on the
input dimension of the agents? policies. When the policy dimension is too low, many agents
share identical policies, and the system suffers from the maladaptive behavior of the agents,
meaning that they prematurely rush to adapt to system changes in bursts [4].
Recently, we have demonstrated that a better system ef?ciency can be attained by introducing diversity [5]. This is done by randomly assigning biases to the initial preference of
policies of the agents, so that agents sharing common policies may not adopt them at the
same time, and maladaptation is reduced. As a result, the population difference between the
majority and minority groups decreases. For typical control tasks such as the distribution of
shared resources, this corresponds to a high system ef?ciency. In contrast to the maladaptive regime, in which agents blindly respond to environmental signals, agent cooperation
becomes increasingly important in the diverse regime. Namely, there are fewer agents ad-
justing their policy perferences at each step of the steady state, but there emerges a more
coordinated pattern of policy adjustment among them. Hence, it is interesting to study the
mechanisms by which they adapt mutually, and their effects on the system ef?ciency.
In this paper, we explain the cooperative mechanisms which appear successively when
the diversity of the agents? preference of policies increases, as recently proposed in [6].
We will provide experimental evidence of these effects, and sketch their analyses which
yield excellent agreement with simulations. While we focus on the population dynamics
of the Minority Game, we expect that the observed cooperative mechanisms are relevant to
reinforcement learning in multi-agent systems more generally.
2
The Minority Game
The Minority Game consists of a population of N agents competing sel?shly to maximize
their individual utility in an environment of limited resources, N being odd [3]. Each agent
makes a decision + or ? at each time step, and the minority group wins. For typical control
tasks such as the resource allocation, the decisions + and ? may represent two alternative
resources, so that less agents utilizing a resource implies more abundance. The decisions
of each agent are prescribed by policies, which are binary functions mapping the history of
the winning bits of the game in the most recent m steps to decisions + or ?. Hence, m is
the memory size. Before the game starts, each agent randomly picks s policies out of the
set of 2D policies with replacement, where D ? 2m is the number of input states.
The long-term goal of an agent is to maximize her cumulative payoff, which is the sum
of the undiscounted payoffs received during the game history. For the decision
P ? i (t) of
agent i at time t (?i (t) = ?1), the payoff is ??i (t)G(A(t)), where A(t) ? i ?i (t)/N ,
and G(A) satis?es the property signG(A) = signA. She tries to achieve her goal by
choosing at each step, out of her s policies, the most successful one so far, and outputing
her decision accordingly. The success of a policy is measured by its cumulative payoff,
updated every step irrespective of whether it is adopted or not. This reinforcement learning
provides an agent with adaptivity. Though we only consider random policies instead of
organized ones, we expect that the model is suf?cient to capture the collective behavior of
large population games. In this paper, we consider a step payoff function, G(A) = signA.
The cumulative payoffs then take integer values. Note that an agent gains in payoff when
she makes a decision opposite to A(t), and loses otherwise, re?ecting the winning of the
minority group.
It is natural to consider systems with diverse preferences of policies [5]. This means that
the initial cumulative payoffs of policies ? (? = 1, . . . , s ? 1) of agent i with respect to
her sth policy have random biases ?i? . Diversity is important in reducing the maladaptive
behavior of the agents, since otherwise the same policy of all agents accumulates the same
payoffs, and would be adopted at the same time. In this paper, we consider the case s = 2,
and the biases are the sums of ?1 randomly drawn R times. In particular, when R is not too
small, the bias distribution approaches a Gaussian distribution with mean 0 and variance R.
The ratio ? ? R/N is referred to as the diversity. For odd R, no two policies have the same
cumulative payoffs throughout the process, and the dynamics is deterministic, resulting in
highly precise simulation results useful for re?ned comparison with theories.
The population averages of the decisions oscillate around 0 at the steady state. Since a
large difference between the majority and minority populations implies inef?cient resource
allocation, the inef?ciency of the system is often measured by the variance ? 2 /N of the
population making decision +, and is given by
?
?
?2
N
? h[A? (t) (t) ? hA? (t) (t)it ]2 it ,
(1)
N
4
where h it denotes time average at the steady state. Its dependence on the diversity is
10
0
2
4
? /N
2
? /N
10
10
?2
10
?3
N = 127
N = 511
N = 2047
N = 8191
N = 32767
10
N = 127
N = 511
N = 2047
?4
10
?6
10
?2
10
2
10
6
?
10
?4
10
2
10
4
10
6
?
Figure 1: (a) The dependence of the variance of the population making decision + on the
diversity at m = 1 and s = 2. Symbols: simulation results averaged over 1,024 samples of
initial conditions. Lines: theory. Dashed-dotted line: scaling prediction. (b) Comparison
between simulation results (symbols), theory with kinetic sampling only (dashed lines),
one-wait approximation (dash-dotted lines), and many-wait approximation (lines).
shown in Fig. 1. Several modes of agent cooperation can be identi?ed, and explained in the
following sections.
3
Statistical Cooperation
For each curve with a given N in Fig. 1(a), and besides the ?rst few data points where
? ? N ?1 and ? 2 /N ? N , the behavior of the variance is dominated by the scaling relation
? 2 /N ? ??1 for ? ? 1. To interpret this result, we describe the macroscopic dynamics of
the system by de?ning the D-dimensional vector A ? (t), which is the sum of the decisions
of all agents responding to history ? of their policies, normalized by N . While only one of
the D components corresponds to the historical state ?? (t) of the system, the augmentation
to D components is necessary to describe the attractor structure and the transient behavior
of the system dynamics.
The key to analysing the system dynamics is the observation that the cumulative payoffs
of all policies displace by exactly the same amount when the game proceeds. Hence for
a given pair of policies, the pro?le of the relative cumulative payoff distribution remains
unchanged, but the peak position shifts with the game dynamics. Let us consider the change
in A? (t) when ? is the historical state ?? (t). We let S?? (?) be the number of agents
holding policies ? and ? (with ? < ?), and the bias of ? with respect to ? is ?. If the
cumulative payoff of policy ? at time t is ?? (t), then the agents holding policies ? and ?
make decisions according to policy ? if ? + ?? (t) ? ?? (t) > 0, and policy ? otherwise.
Hence ? + ?? (t) ? ?? (t) is referred to as the preference of ? with respect to ?. At time
t, the cumulative payoff of policy ? changes from ?? (t) to ?? (t) ? ??? signA? (t), where
??? is the decision of policy ? at state ?. Only the ?ckle agents, that is, those agents with
preferences on the verge of switching signs, contribute to the change in A? (t), namely,
? + ?? (t) ? ?? (t) = ?1 and ??? ? ??? = ?2signA? (t). Hence we have
A? (t + 1) ? A? (t) =
?signA? (t)
2 X X
S?? (r ? ?? (t) + ?? (t))
N
r=?1
?<?
??(??? ? ??? ? 2rsignA? (t))
(2)
where ?(n) = 1 if n = 0, and 0 otherwise. In the region where D ? ln N , we have
Prob =
1
8
1
8
A
?
Q
0.6
P
Prob =
3
8
A
R
Prob =
3
8
+
average number of agents
Prob =
0.4
0.2
0.0
?3001
?1
3001
preference
Figure 2: (a) The attractor in the Minority Game with m = 1, following the period-4
sequence of P-Q-R-Q in the phase space of A+ and A? . There are 4 approaches to the
attractor indicated by the arrows, and the respective probabilities are obtained by considering the detailed dynamics from the different initial positions and states. (b) Experimental
evidence of the kinetic sampling effect: steady-state preference dependence of the average number of agents holding the identity policy and its complement, immediately before
state Q enters state R, at ? = N = 1, 023 and averaged over 100,000 samples of initial
conditions.
S?? (?) ? 1, and Eq. (2) is self-averaging. Following the derivation in [5], we arrive at
r
2
?
?
?
A (t + 1) = A (t) ? signA (t)
?(? ? ?? (t)).
(3)
?R
Equation (3) shows that the dynamics proceeds in the direction
which reduces the magnip
tude of the population vector, each time by a step of size 2/?R. At the steady state, each
component oscillates between positive and negative, as shown in the example of m = 1
in Fig. 2(a). Due to the maladaptive nature of the dynamics, it never reaches p
the zero
value. As a result, each state is con?ned in a D-dimensional hypercube of size 2/?R,
irrespective of the initial position of the population vector. This con?nement enables us
to compute the variance of the decisions, given by ? 2 /N = f (?)/2??, where f (?) is a
smooth function of ?, which approaches (1 ? 1/4D)/3 for ? ? 1. The physical picture of
this scaling relation comes from the broadening of the preference distribution due to bias
diversity. The fraction of ?ckle agents at every time step consists of those who have ?1
preferences, which scales as the height of the bias?distribution near its center. Since
the
?
distribution is a Gaussian with standard deviation R, the step sizes scale as 1/ R, and
variances ? 2 /N as ??1 . The scaling relation shows that agent cooperation in this regime is
described at the level of statistical distributions of policy preferences, since
? the number of
agents making an adaptive move at each step is suf?ciently numerous (? N ).
4
Kinetic Sampling
As shown in Fig. 1(a), ? 2 /N deviates above the scaling with ??1 when ? ? N . To consider
the origin of this deviation, we focus in Fig. 2(b) on how the average number of agents, who
hold the identity policy with ??? = ? and its complementary policy ??? = ??, depends on
the preference ?+?? ??? , when the system reaches the steady state in games with m = 1.
Since the preferences are time dependent, we sample their frequencies at a ?xed time, say,
immediately before the state changes from Q to R in Fig. 2(a). One would expect that the
bias distribution is reproduced. However, we ?nd that a peak exists at ? + ? ? ? ?? = ?1.
This value of the preference corresponds to that of the attractor step from Q to R when at
state ?, decision + loses and decision ? wins, and ? + ?? ? ?? changes from ?1 to +1.
The peak at the attractor step shows that its average size is self-organized to be larger than
those of the transient steps described by the background distribution.
This effect that favors the cooperation of larger clusters of agents is referred to as the
kinetic sampling effect. When ? ? N , A? (t + 1) ? A? (t) scales as N ?1 and is no longer
self-averaging. Rather, Eq. (2) shows that it is equal to 2/N times
? the number of ?ckle
agents atptime t, which is Poisson distributed with a mean of N/ 2?R = ?/2, where
? ? N 2/?R is the average step size. However, since the attractor is formed by steps
which reverse the sign of A? , the average step size in the attractor is larger than that in
the transient state, because a long jump in the vicinity of the attractor is more likely to get
trapped.
To describe this effect, we consider the probability Patt (?A) of step sizes ?A in the
attractor (with ?A? > 0 for all ?). Assuming
Pthat all states of the phase space are equally
likely to be accessed, we have Patt (?A) = A Patt (?A, A), where Patt (?A, A) is the
probability of ?nding the position A with displacement ?A in the attractor. Consider the
example of m = 1, where there is only one step Q
along each axis A? . The sign reversal
condition implies that Patt (?A, A) ? PPoi (?A) ? ?[?A? (A? + ?A? )], where ?(x)
is the step function of x, Q
and PPoi (?A) is the Poisson distribution of step sizes, yielding
Patt (?A) ? PPoi (?A) ? ?A? . We note that the extra factors of ?A? favor larger step
sizes. Thus, the attractor averages h(?A? )2 iatt are given by
h(?A? )2 iatt =
h(?A? )2 ?A+ ?A? iPoi
.
h?A+ ?A? iPoi
(4)
There are agents who contribute to both ?A+ and ?A? , giving rise to their correlations.
In Eq. (2), the strategies of the agents contributing to ?A+ and ?A? satisfy ??+ ? ??+ =
?2r and ??? ? ??? = 2r respectively. Among the agents contributing to ?A+ , the extra
requirement of ??? ? ??? = 2r implies that an average of 1/4 of them also contribute to
?A? . Hence, the number of agents contributing to both steps is a Poisson variable with
mean ?/8, and those exclusive to the individual steps are Poisson variables with mean
3?/8. This yields, for example,
?a0 e? 3?
? 3? ?a+ e? 3?
? 3? ?a?
P
?? ?
8
8
h?A+ ?A? iPoi = N42 a0 ,a+ ,a? ea08! ?
8
a+ !
8
a? !
8
(a0 + a+ )(a0 + a? ).
(5)
Together with similar expressions of the numerator in Eq. (4), we obtain
h(?A? )2 iatt =
2?3 + 15?2 + 20? + 4
.
N 2 (2? + 1)
(6)
The attractor states are given by A? = m? /N and m? /N ? ?A? , where m? =
1, 3, . . . , N ?A? ? 1. This yields a variance of
?2
7h(N ?A+ )2 iatt + 7h(N ?A? )2 iatt ? 8
=
,
N
192N
which gives, on combining with Eq. (6),
?2
14?3 + 105?2 + 132? + 24
=
.
N
96N (2? + 1)
(7)
(8)
When the diversity is low, ? ? 1, and Eq. (8) reduces to ? 2 /N = 7/48??, agreeing with
the scaling result of the previous section. When ? ? N , Eq. (8) has excellent agreement
with simulation results, which signi?cantly deviate above the scaling relation.
5
Waiting Effect
As shown in Fig. 1(b), ? 2 /N further deviates above the predictions of kinetic sampling
when ? ? N . To study the origin of this effect, we consider the example of m = 1. As
shown in Fig. 2(a), the attractor consists of both hops along the A? axes. Analysis shows
that only those agents holding the identity policy and its complement can complete both
hops after they have adjusted their preferences to ? + ?? ? ?? = ?1. Since there are
fewer and fewer ?ckle agents in the limit ? ? N , one would expect that a single agent of
this type would dominate the game dynamics, and ? 2 /N would approach 0.25/N , as also
predicted by Eq. (8).
However, attractors having 2 ?ckle agents are about 10 times more common in the extremely diverse limit. As illustrated in Fig. 3(a) for a typical case, one of the two agents
?rst arrives at the status of ?1 preference of her policies and stay there waiting. Meanwhile, the preference of the second agent is steadily reduced. Once she has arrived at the
status of ?1 preference of her policies, both agents can then cooperate to complete the
dynamics of the attractor. In this example, both agents do not belong to the correct type
that can complete the dynamics alone, but waiting is crucial for them to complete the hops
in the attractor, even though one would expect that the probability of ?nding more than one
?ckle agents at a time step is drastically less than that for one. Thus, the composition of
the group of ?ckle agents is self-organized through this waiting effect, and consequently
the step sizes and variance increase above those predicted by kinetic sampling.
The analysis of the waiting effect is lengthy. Here the agents are so diverse that the average
step size is approaching 0. At each state in the phase space, the system remains stationary
for many time steps, waiting for some agent to reduce the magnitude of her preference until
policy switching can take place. For illustration, we sketch the approximation of including
up to one wait. As shown in Fig. 2(a), the attractor may be approached from the arm (P
or R) or from the corner (Q). Consider the case of the state approaching from P, waiting
up to k times at Q to move to R, and ending the transient dynamics thereafter. Then the
cumulative payoffs of a policy ? can be written as ?? + ??+ at P, ?? , . . . , ?? ? k??? at Q
and, in the attractor of period 4, repeating the sequence of ?? ? k??? ? ??? at R, ?? ? k???
at Q, ?? ? k??? + ??+ at P, and ?? ? k??? at Q. The
P movement of the cumulative payoffs
can be conveniently represented by writing ?? = ? k ? ??? , where k ? denotes the number
of wins minus losses of decision 1 at state ? in the game history. For m = 1, these steps
are plotted in the space of k + and k ? in Fig. 3(b).
The size of each step is 2/N times the number of ?ckle agents at that step, which is Poisson
distributed with average ?/2. The average numbers of agents appearing simultaneously in
different steps positioned along the directions k + ?k ? = constant and k ? = constant are,
respectively, ?/8 and ?/4, and 0 for other directions. Thus, the average number of agents
common in the pairs of steps {PQ, QQ1 }, {QQk , QP}, {QP, QR}, {PQ, QP} are ?/8,
?/8, ?/8 and ?/4 respectively. The rest of the combinations of steps are uncorrelated.
The number of agents involved in the steps are described in Table 1.
The variance of the step sizes is given by
X
1
h [(?A+ )2 +(?A? )2 ]iatt =
Pj
2
j
?P
1
+ 2
i=0,1 h 2 [(?A )
P
+ (?A? )2 ]?A+ ?A? ii,j
i=0,1 h?A
+ ?A? i
i,j
!
,
(9)
where j = arm or corner. The variance of decisions can then be obtained from Eq. (7). For
illustration, we consider the derivation of the Poisson average h?A+ ?A? i for one-wait
2000
k
2nd agent
?
QQ 1
PQ
Preference
1000
0
QQ r
1st agent
?1000
QQ k
?2000
QP
QR
0
400
800
1200
Time
k
+
Figure 3: (a) Experimental evidence of the waiting effect: a typical example of the evolution of the preference of the 2 agents switching policies at the attractor in a game with
m = 1, N = 127, and R = 224 ?1. The system converges to the attractor at t = 1, 086. (b)
The space of k + and k ? describing the movement of the cumulative payoffs in the game
with m = 1. Thick arrows: non-vanishing steps. Thin arrows: waiting steps. Thick double
arrows: attractor steps. The dashed lines link those steps that share common agents.
arm approach. Noting that the only non-vanishing steps are PQ, QR and QP, we obtain
P?
h?A+ ?A? i1,arm = N42 k=1 h[1 ? ?(ai )?(aturn,1 )?(acum )]?(aturn,1 )
Qk
r=1 ?(await,r )?(aturn,2 )(a? + a0 )(acum + aturn,2 + a0 )iPoi
n ? h ? ?2
i
io
h ? ?2
?
?
?
? 7?
8
?
e
.
(10)
= N42 1? ? e? 2 12 ?
+
4
+
8
8
8
8
1?e
2
We note that the number a0 accounts for the agents who contribute to both steps in the
attractor, and thus can complete the attractor dynamics alone in the extremely diverse limit.
On the other hand, the number acum arises from the ?rst step PQ arriving at Q. Once
present, it will appear in the attractor step QP, irrespective of the duration k of the wait at
Q. These acum agents can wait to complete the attractor dynamics together with the a?
agents who contribute independently to the step from Q to R, as well as the a0 agents who
contribute to both attractor steps. As a result, the average step size increases due to this
waiting effect. In the former case, cooperation between individual types of agents becomes
indispensable in reaching the steady state behavior.
Other Poisson averages in Eq. (9) can be derived similarly. As shown in Fig. 1(b), the waiting effect causes the variance to increase beyond the kinetic sampling prediction, agreeing
with the trend of the simulation results. In particular, the variance approaches 0.34/N in
the extremely diverse limit, signi?cantly greater than the limit of 0.25/N in the absence of
waiting effects. Further approximation including multiple waiting steps results in the theoretical curves with excellent agreement with the simulation results, as shown in Fig. 1(b).
In the extremely diverse limit, the theoretical predictions approach 0.42/N , very close to
the simulation result of 0.43/N .
6
Conclusion
We have studied the dynamical mechanisms of cooperation, which emerges automatically
in a multi-agent system with adaptive agents competing sel?shly for ?nite resources. At
low diversity, agent cooperation proceeds at the statistical level, resulting in the scaling
relation of the variance with diversity. At high diversity, when kinetic sampling becomes
Label
PQ
QQ1
QQr
QQk
QR
QP
Table 1: The number of ?ckle agents in the steps of one wait.
Steps
No. of agents
Poisson averages
?? + ??+ ? ?? ai + aturn,1 + acum hai i = ?/8, haturn,1 i = ?/8,
hacum i = ?/4.
?? ? ?? ? ??? await,1 + aturn,1
hawait,1 i = 3?/8.
?? ? (r ? 1)???
? ?? ? r???
?? ? (k ? 1)???
? ?? ? k???
?? ? k??? ?
?? ? (k + 1)???
?? ? k??? ?
?? ? k??? + ??+
await,r
await,k + aturn,2
a? + a 0
acum + aturn,2 + a0
hawait,r i = ?/2,
(2 ? r ? k ? 1).
hawait,k i = 3?/8,
haturn,2 i = ?/8.
ha? i = 3?/8,
ha0 i = ?/8.
signi?cant, we ?nd that the attractor dynamics favors the cooperation of larger clusters
of agents. In extremely diverse systems, we further discover a waiting mechanism, when
agents who are unable to complete the attractor dynamics alone wait for other agents to
collaborate with them. When waiting is present, cooperation between individual types
of agents becomes indispensable in reaching the steady state behavior. Together, these
mechanisms yield theoretical predictions of the population variance in excellent agreement
with simulations over nine decades of data.
We expect that the observed mechanisms of agent cooperation can be found in reinforcement learning of multi-agent systems in general, due to their generic nature. The mechanisms of statistical cooperation, kinetic sampling and waiting illustrate the importance of
dynamical considerations in describing the system behavior, and the capability of multiagent systems to self-organize in their collective dynamics. In particular, it is interesting to
note that given enough waiting time, agents with limited abilities can cooperate to achieve
dynamics unachievable by individuals. This is relevant to evolutionary approaches to multiagent control, since it allows limited changes to accumulate into bigger improvements.
Acknowledgments
We thank C. H. Yeung, Y. S. Ting and B. H. Wang for fruitful discussions. This
work is supported by the Research Grant Council of Hong Kong (HKUST6153/01P,
HKUST6062/02P) and DAG04/05.SC25.
References
[1] G. Wei? and S. Sen, Adaption and Learning in Multi-agent Systems, Lecture Notes in Computer
Science 1042 (Springer, Berlin, 1995).
[2] E. Rasmusen, Games and Information (Basil Blackwell, Oxford, 2001).
[3] D. Challet and Y. C. Zhang, Emergence of Cooperation and Organization in an Evolutionary
Game, Physica A 246, pp. 407-418 (1997).
[4] R. Savit, R. Manuca, and R. Riolo, Adaptive Competition, Market Ef?ciency, and Phase Transitions, Phys. Rev. Lett. 82, pp. 2203-2206 (1999).
[5] K. Y. M. Wong, S. W. Lim, and P. Luo, Diversity and Adaptation in Large Population Games,
Int. J. Mod. Phys. B 18, 2422-2431 (2004).
[6] K. Y. M. Wong, S. W. Lim, and Z. Gao, Dynamical Mehanisms of Adaptation in Multi-agent
Systems, Phys. Rev. E 70, 025103(R) (2004).
| 2687 |@word kong:3 version:1 nd:3 simulation:10 pick:1 thereby:1 minus:1 initial:7 luo:1 assigning:1 ust:1 written:1 cant:1 enables:1 displace:1 alone:3 stationary:1 fewer:3 accordingly:1 vanishing:2 provides:1 contribute:6 preference:23 accessed:1 zhang:1 height:1 burst:1 along:3 consists:3 overhead:1 market:1 behavior:12 multi:8 automatically:1 considering:1 becomes:7 discover:1 xed:1 every:2 exactly:1 oscillates:1 control:4 grant:1 appear:2 organize:1 before:3 positive:1 limit:6 io:1 switching:3 accumulates:1 oxford:1 china:1 studied:1 limited:4 averaged:2 pthat:1 acknowledgment:1 nite:1 displacement:1 wait:8 get:1 close:1 writing:1 wong:3 fruitful:1 deterministic:1 demonstrated:1 center:1 duration:1 independently:1 immediately:2 utilizing:1 dominate:1 population:14 updated:1 qq:3 origin:2 agreement:5 trend:1 maladaptive:4 cooperative:2 observed:2 enters:1 capture:1 await:4 wang:1 region:1 decrease:1 movement:2 environment:2 nash:1 dynamic:21 ha0:1 represented:1 derivation:2 describe:3 approached:2 choosing:1 whose:1 larger:5 say:1 otherwise:4 favor:3 ability:1 emergence:1 reproduced:1 advantage:1 sequence:2 analytical:1 sen:1 adaptation:2 relevant:2 inef:2 combining:1 achieve:2 competition:1 scalability:1 qr:4 rst:3 cluster:2 undiscounted:1 requirement:1 double:1 converges:1 illustrate:1 measured:2 odd:2 received:1 eq:10 predicted:2 involves:1 implies:4 come:1 signi:3 direction:3 ning:1 thick:2 correct:1 transient:4 adjusted:1 physica:1 hold:2 around:1 equilibrium:1 mapping:1 adopt:1 label:1 council:1 gaussian:2 rather:1 reaching:2 sel:2 ax:1 focus:2 derived:1 she:3 improvement:1 hk:1 contrast:1 dependent:1 a0:9 her:8 relation:5 i1:1 among:2 diversi:1 equal:1 once:2 never:1 having:1 sampling:9 hop:3 identical:1 thin:1 intelligent:1 few:1 randomly:3 simultaneously:2 individual:5 phase:4 replacement:1 attractor:28 organization:1 interest:1 satis:1 highly:2 arrives:1 yielding:1 light:1 necessary:1 respective:1 re:2 plotted:1 rush:1 theoretical:3 introducing:1 deviation:2 successful:1 too:2 adaptively:1 st:1 peak:3 stay:1 cantly:2 michael:1 together:3 augmentation:1 successively:1 corner:2 verge:1 strive:1 account:1 diversity:14 de:1 int:1 coordinated:1 satisfy:1 depends:2 ad:1 try:1 start:1 capability:1 formed:1 variance:14 who:7 qk:1 yield:5 history:4 explain:2 reach:2 suffers:1 phys:3 sharing:1 ed:2 lengthy:1 frequency:1 steadily:1 involved:1 pp:2 con:2 gain:1 lim:3 emerges:2 organized:3 positioned:1 attained:1 wei:1 done:1 though:2 correlation:1 until:1 sketch:2 hand:1 mode:1 indicated:1 effect:14 normalized:1 evolution:1 hence:6 vicinity:1 former:1 illustrated:1 game:22 during:1 self:5 numerator:1 steady:10 hong:3 outputing:1 arrived:1 complete:7 pro:1 cooperate:2 meaning:1 consideration:1 ef:4 recently:2 common:4 physical:1 qp:7 belong:1 interpret:1 accumulate:1 composition:1 ai:2 collaborate:1 similarly:1 pq:6 longer:1 showed:1 recent:1 reverse:1 indispensable:2 binary:1 success:1 greater:1 speci:1 maximize:2 period:2 signal:1 dashed:3 ii:1 multiple:1 reduces:2 smooth:1 adapt:3 long:2 equally:1 bigger:1 prediction:5 variant:1 heterogeneous:1 poisson:8 blindly:1 yeung:1 represent:1 background:1 macroscopic:1 crucial:1 extra:2 rest:1 mod:1 integer:1 ciently:1 near:1 noting:1 enough:1 competing:3 opposite:1 approaching:2 reduce:1 shift:1 whether:1 expression:1 utility:1 nement:1 nine:2 oscillate:1 cause:1 generally:1 useful:1 detailed:1 tude:1 amount:1 repeating:1 reduced:2 dotted:2 sign:3 trapped:1 diverse:9 patt:6 waiting:17 group:5 key:1 thereafter:1 basil:1 drawn:1 pj:1 fraction:1 sum:3 compete:1 prob:4 respond:1 arrive:1 throughout:1 place:1 decision:20 scaling:8 bit:1 dash:1 unachievable:1 dominated:1 qqk:2 prescribed:1 ecting:1 extremely:5 ned:2 according:1 combination:1 increasingly:2 agreeing:2 sth:1 rev:2 making:4 explained:1 ln:1 resource:9 mutually:1 remains:2 equation:1 describing:2 mechanism:9 reversal:1 adopted:2 generic:1 appearing:1 alternative:1 robustness:1 denotes:2 responding:1 cally:1 giving:1 ting:1 hypercube:1 unchanged:1 move:2 strategy:1 dependence:3 exclusive:1 hai:1 evolutionary:2 win:3 link:1 unable:1 thank:1 berlin:1 majority:3 minority:12 assuming:1 besides:1 illustration:2 ratio:1 holding:4 negative:1 rise:1 collective:3 policy:42 observation:1 payoff:18 communication:1 precise:1 prematurely:1 interacting:1 introduced:1 complement:2 namely:2 pair:2 blackwell:1 identi:1 beyond:1 proceeds:3 parallelism:1 pattern:1 dynamical:3 regime:3 including:2 memory:1 natural:1 ipoi:4 arm:4 technology:1 numerous:1 picture:1 nding:2 irrespective:3 axis:1 deviate:3 contributing:3 relative:1 loss:1 expect:6 multiagent:2 lecture:1 adaptivity:1 interesting:3 suf:2 allocation:2 agent:90 signa:6 uncorrelated:1 share:2 balancing:1 cooperation:16 supported:1 arriving:1 drastically:1 bias:9 riolo:1 distributed:3 curve:2 dimension:2 lett:1 ending:1 cumulative:13 transition:1 reinforcement:4 adaptive:4 jump:1 historical:2 far:1 status:2 decade:2 table:2 nature:2 broadening:1 excellent:5 complex:1 meanwhile:1 arrow:4 complementary:1 fig:13 referred:3 cient:2 position:4 ciency:5 winning:2 ppoi:3 abundance:1 load:1 phkywong:1 symbol:2 striving:1 evidence:4 exists:1 adding:1 importance:1 magnitude:1 likely:2 gao:2 conveniently:1 adjustment:1 springer:1 corresponds:3 loses:2 environmental:1 adaption:1 kinetic:9 goal:2 identity:3 consequently:1 shared:1 absence:1 change:7 analysing:1 typical:4 reducing:1 averaging:2 qq1:2 experimental:3 e:1 n42:3 arises:1 |
1,858 | 2,688 | Modelling Uncertainty in the Game of Go
David H. Stern
Department of Physics
Cambridge University
[email protected]
Thore Graepel
Microsoft Research
Cambridge, U.K.
[email protected]
David J. C. MacKay
Department of Physics
Cambridge University
[email protected]
Abstract
Go is an ancient oriental game whose complexity has defeated attempts to automate it. We suggest using probability in a Bayesian
sense to model the uncertainty arising from the vast complexity
of the game tree. We present a simple conditional Markov random field model for predicting the pointwise territory outcome of
a game. The topology of the model reflects the spatial structure of
the Go board. We describe a version of the Swendsen-Wang process for sampling from the model during learning and apply loopy
belief propagation for rapid inference and prediction. The model
is trained on several hundred records of professional games. Our
experimental results indicate that the model successfully learns to
predict territory despite its simplicity.
1
Introduction
The game of Go originated in China over 4000 years ago. Its rules are simple (See
www.gobase.org for an introduction). Two players, Black and White, take turns
to place stones on the intersections of an N ? N grid (usually N = 19 but smaller
boards are in use as well). All the stones of each player are identical. Players place
their stones in order to create territory by occupying or surrounding areas of the
board. The player with the most territory at the end of the game is the winner. A
stone is captured if it has been completely surrounded (in the horizontal and vertical
directions) by stones of the opponent?s colour. Stones in a contiguous ?chain? have
the common fate property: they are captured all together or not at all [1].
The game that emerges from these simple rules has a complexity that defeats attempts to apply minimax search. The best Go programs play only at the level of
weak amateur Go players and Go is therefore considered to be a serious AI challenge
not unlike Chess in the 1960s. There are two main reasons for this state of affairs:
firstly, the high branching factor of Go (typically 200 to 300 potential moves per
position) prevents the expansion of a game tree to any useful depth. Secondly, it
is difficult to produce an evaluation function for Go positions. A Go stone has no
intrinsic value; its value is determined by its relationships with other stones. Go
players evaluate positions using visual pattern recognition and qualitative intuitions
which are difficult to formalise.
Most Go programs rely on a large amount of hand-tailored rules and expert knowl-
edge [2]. Some machine learning techniques have been applied to Go with limited
success. Schraudolph, Dayan and Sejnowski [3] trained a multi-layer perceptron to
evaluate board positions by temporal difference learning. Enzenberger [4] improved
on this by structuring the topologies of his neural networks according to the relationships between stones on the board. Graepel et al. [1] made use of the common
fate property of chains to construct an efficient graph-based representation of the
board. They trained a Support Vector Machine to use this representation to solve
Go problems.
Our starting point is the uncertainty about the future course of the game that
arises from the vast complexity of the game tree. We propose to explicitly model
this uncertainty using probability in a Bayesian sense. The Japanese have a word,
aji, much used by Go players. Taken literally it means ?taste?. Taste lingers, and
likewise the influence of a Go stone lingers (even if it appears weak or dead) because
of the uncertainty of the effect it may have in the future. We use a probabilistic
model that takes the current board position and predicts for every intersection of
the board if it will be Black or White territory. Given such a model the score of
the game can be predicted and hence an evaluation function produced. The model
is a conditional Markov random field [5] which incorporates the spatial structure of
the Go board.
2
Models for Predicting Territory
Consider the Go board as an undirected Graph G = (N , E) with N = Nx ?Ny nodes
n ? N representing vertices on the board and edges e ? E connecting vertically
and horizontally neighbouring points. We can denote a position as the vector c ?
{Black, White, Empty}N for cn = c(n) and similarly the final territory outcome
of the game as s ? {+1, ?1}N for sn = s(n). For convenience we score from
the point of view of Black so elements of s representing Black territory are valued
+1 and elements representing white territory are valued ?1. Go players will note
that we are adopting the Chinese method of scoring empty as well as occupied
intersections. The distribution we wish to model is P (s|c), that is, the distribution
over final territory outcomes given the current position. Such a model would be
useful for several reasons.
? Most importantly, the detailed outcomes provide us with a simple
P evaluation function for Go positions by the expected score, u(c) := h i si iP (s|c) .
An alternative (and probably better) evaluation function is givenPby the
probability of winning which takes the form P (Black wins) = P ( i si >
komi), where komi refers to the winning threshold for Black.
? Connectivity of stones is vital because stones can draw strength from other
stones. Connectivity could be measured by the correlation between nodes
under the distribution P (s|c). This would allow us to segment the board
into ?groups? of stones to reduce complexity.
? It would also be useful to observe cases where we have an anti-correlation
between nodes in the territory prediction. Japanese refer to such cases as
miai in which only one of two desired results can be achieved at the expense
of the other - a consequence of moving in turns.
? The fate of a group of Go stones could be estimated from the distribution
P (s|c) by marginalising out the nodes not involved.
The way stones exert long range influence can be considered recursive. A stone
influences its neighbours, who influence their neighbours and so on. A simple model
which exploits this idea is to consider the Go board itself as an undirected graphical
model in the form of a Conditional Random Field (CRF) [5]. We factorize the
distribution as
?
?
Y
X
1
1
P (s|c) =
?f (sf , cf , ? f ) =
exp ?
log(?f (sf , cf , ? f ))? .
Z(c, ?)
Z(c, ?)
f ?F
f ?F
(1)
The simplest form of this model has one factor for each pair of neighbouring nodes
i, j so ?f (sf , cf , ? f ) = ?f (si , sj , ci , cj , ? f ).
Boltzmann5 For our first model we decompose the factors into ?coupling? terms
and ?external field? terms as follows:
?
?
X
1
P (s|c) =
exp ?
{w(ci , cj )si sj + h(ci )si + h(cj )sj }?
(2)
Z(c, ?)
(i,j)?F
This gives a Boltzmann machine whose connections have the grid topology of the
board. The couplings between territory-outcome nodes depend on the current board
position local to those nodes and the external field at each node is determined
by the state of the board at that location. We assume that Go positions with
their associated territory positions are symmetric with respect to colour reversal so
?f (si , sj , ci , cj , ? f ) = ?f (?si , ?sj , ?ci , ?cj , ? f ). Pairwise connections are also invariant to direction reversal so ?f (si , sj , ci , cj , ? f ) = ?f (sj , si , cj , ci , ? f ). It follows
that the model described in 2 can be specified by just five parameters:
?
?
?
?
?
wchains = w(Black, Black) = w(White, White),
winter?chain = w(Black, White) = w(White, Black),
wchain?empty = w(Empty, White) = w(Empty, Black),
wempty = w(Empty, Empty),
hstones = h(Black) = ?h(White),
and h(empty) is set to zero by symmetry. We will refer to this model as Boltzmann5.
This simple model is interesting because all these parameters are readily interpreted.
For example we would expect wchains to take on a large positive value since chains
have common fate.
BoltzmannLiberties A feature that has particular utility for evaluating Go positions is the number of liberties associated with a chain of stones. A liberty of a
chain is an empty vertex adjacent to it. The number of liberties indicates a chain?s
safety because the opponent would have to occupy all the liberties to capture the
chain. Our second model takes this information into account:
?
?
X
1
P (s|c) =
exp ?
w(ci , cj , si , sj , li , lj )? ,
(3)
Z(c, ?)
(i,j)?F
where li is element i of a vector l ? {+1, +2, +3, 4 or more}N the liberty count of
each vertex on the Go board. A group with four or more liberties is considered
relatively safe. Again we can apply symmetry arguments and end up with 78
parameters. We will refer to this model as BoltzmannLiberties.
We trained the two models using board positions from a database of 22,000 games
between expert Go players1 . The territory outcomes of a subset of these games
1
The GoGoD database, April 2003. URL:http://www.gogod.demon.co.uk
(a) Gibbs Sampling
(b) Swendsen Wang
Figure 1: Comparing ordinary Gibbs with Swendsen Wang sampling for Boltzmann5. Shown are the differences between the running averages and the exact
marginals for each of the 361 nodes plotted as a function of the number of wholeboard samples.
were determined using the Go program GnuGo2 to analyse their final positions.
Each training example comprised a board position c, with its associated territory
outcome s. Training was performed by maximising the likelihood ln P (s 0 |c) using
gradient descent. In order to calculate the likelihood it is necessary to perform
inference to obtain the marginal expectations of the potentials.
3
Inference Methods
It is possible to perform exact inference on the model by variable elimination [6].
Eliminating nodes one diagonal at a time gave an efficient computation. The cost
of exact inference was still too high for general use but it was used to compare other
inference methods.
Sampling The standard method for sampling from a Boltzmann machine is to use
Gibbs sampling where each node is updated one at a time, conditional on the others.
However, Gibbs sampling mixes slowly for spin systems with strong correlations.
A generalisation of the Swendsen-Wang process [7] alleviates this problem. The
original Swendsen-Wang algorithm samples from a ferromagnetic Ising model with
no external field by adding an additional set of ?bond? nodes d, one attached to
each factor (edge) in the original graph. Each of these nodes can either be in the
state ?bond? or ?no bond?. The new factor potentials ?f (sf , cf , df , ? f ) are chosen
such that if a bond exists between a pair of spins then they are forced to be in
the same state. Conditional on the bonds, each cluster has an equal probability
of having all its spins in the ?up? state or all in the ?down? state. The algorithm
samples from P (s|d, c, ?) and P (d|s, c, ?) in turn (flipping clusters and forming
bonds respectively). It can be generalised to models with arbitrary couplings and
biases [7, 8]. The new factor potentials ?f (sf , cf , df , ? f ) have the following effect:
if the coupling is positive then when the d node is in the ?bond? state it forces the
two spins to be in the same state; if the coupling is negative the ?bond? state forces
the two spins to be opposite. The probability of each cluster being in each state
depends on the sum of the biases involved. Figure 1 shows that the mixing rate
of the sampling process is improved by using Swendsen-Wang allowing us to find
accurate marginals for a single position in a couple of seconds.
2
URL:http://www.gnu.org/software/gnugo/gnugo.html
Loopy Belief Propagation In order to perform very rapid (approximate) inference we used the loopy belief propagation (BP) algorithm [9] and the results are
examined in Section 4. This algorithm is similar to an influence function [10], as
often used by Go programmers to segment the board into Black and White territory
and for this reason is laid out below.
For each board vertex j ? N , create a data structure called a node containing:
1. A(j), the set of nodes corresponding to the neighbours of vertex j,
2. a set of new messages mnew
ij (sj ) ? Mnew , one for each i ? A(j),
3. a set of old messages mold
ij (sj ) ? Mold , one for each i ? A(j),
4. a belief bj (sj ).
repeat
for all j ? N do
for all i ? A(j) do
for all sj ? {Black, White} do
let variable SUM := 0,
for all si ? {Black, White} do
SUM := SUM + ?(i,j) (si , sj )
Q
mold
qi (si ),
q?A(i)\j
end for
mnew
ij (sj ) := SUM,
end for
end for
end for
for all messages, mnew
xy (sy ) ? Mnew do
new
old
mxy (sy ) := ?mxy (sy ) + (1 ? ?)mnew
xy (sy ),
end for
until completed I iterations (typically I=50)
Belief Update:
for all j ? N do
for all sj ? {Black,
White} do
Q
bj (sj ) :=
mnew
qj (sj )
q?A(j)
end for
end for
Here, ? (typically 0.5), damps any oscillations. ?(i,j) (si , sj ) is the factor potential (see (1)) and in the case of Boltzmann5 takes on the form ?(i,j) (si , sj ) =
exp (w(ci , cj )si sj + h(ci )si + h(cj )sj ). Now the probability of each vertex being
Black or White territory is found by normalising the beliefs at each node. For
example P (sj = Black) = bj (Black)/Z where Z = bj (Black) + bj (White). The
accuracy of the loopy BP approximation appears to be improved by using it during
the parameter learning stage in cases where it is to be used in evaluation.
4
Results for Territory Prediction
Some Learnt Parameters Here are some parameters learnt for the Boltzmann5
model (2). This model was trained on 290 positions from expert Go games at move
80. Training was performed by maximum likelihood as described in Section 2.
(a) Boltzmann5 (Exact)
(b) Boltzmann5 (Loopy BP)
Figure 2: Comparing territory predictions for a Go position from a professional
game at move 90. The circles represent stones. The small black and white squares
at each vertex represent the average territory prediction at that vertex, from ?1
(maximum white square) to +1 (maximum black square).
? hstones = 0.265
? wempty = 0.427
? wchain?empty = 0.442
? wchains = 2.74
? winter?chain = 0.521
The values of these parameters can be interpreted. For example wchains corresponds
to the correlation between the likely territory
outcome of two adjacent vertices in a chain of
connected stones. The high value of this parameter derives from the ?common fate? property of chains as described in Section 1.
Interestingly, the value of the parameter wempty (corresponding to the coupling
between territory predictions of neighbouring vertices in empty space) is 0.427 which
is close to the critical coupling for an Ising model, 0.441.
Territory Predictions Figure 2 gives examples of territory predictions generated
by Boltzmann5. In comparison, Figure 3 shows the prediction of BoltzmannLiberties
and a territory prediction from The Many Faces of Go [2]. Go players confirm that
the territory predictions produced by the models are reasonable, even around loose
groups of Black and White stones. Compare Figures 2 (a) and 3 (a); when liberty
counts are included as features, the model can more confidently identify which of the
two small chains competing in the bottom right of the board is dead. Comparing
Figure 2 (a) and (b) Loopy BP appears to give over-confident predictions in the top
right of the board where few stones are present. However, it is a good approximation
where many stones are present (bottom left).
Comparing Models and Inference Methods Figure 4 shows cross-entropies
between model territory predictions and true final territory outcomes for a dataset
of expert games. As we progress through a game, predictions become more accurate
(not surprising) but the spread of the accuracy increases, possibly due to incorrect
assessment of the life-and-death status of groups. Swendsen-Wang performs better
than Loopy BP, which may suffer from its over-confidence. BoltzmannLiberties
performs better than Boltzmann5 (when using Swendsen-Wang) the difference in
(a) BoltzmannLiberties (Exact)
(b) Many Faces of Go
Figure 3: Diagram (a) is produced by exact inference (training was also by Loopy
BP). Diagram (b) shows the territory predicted by The Many Faces of Go (MFG)
[2]. MFG uses of a rule-based expert system and its prediction for each vertex has
three possible values: ?White?, ?Black? or ?unknown/neutral?.
performance increasing later in the game when liberty counts become more useful.
5
Modelling Move Selection
In order to produce a Go playing program we are interested in modelling the selection of moves. A measure of performance of such a model is the likelihood it assigns
to professional moves as measured by
X X
log P (move|model).
(4)
games moves
We can obtain a probability over moves by choosing a Gibbs distribution with the
negative energy replaced by the evaluation function,
0
P (move|model, w) =
e?u(c ,w)
Z(w)
(5)
where u(c0 , w) is an evaluation function evaluated at the board position c0 resulting
from a given move. The inverse temperature parameter ? determines the degree to
which the move made depends on its evaluation. The territory predictions from the
models Boltzmann5 and BoltzmannLiberties can be combined with the evaluation
function of Section 2 to produce position evaluators.
6
Conclusions
We have presented a probabilistic framework for modelling uncertainty in the game
of Go. A simple model which incorporates the spatial structure of a board position
can perform well at predicting the territory outcomes of Go games. The models
described here could be improved by extracting more features from board positions
and increasing the size of the factors (see (1)).
1.0
Cross Entropy
0.0
0.5
1.0
0.5
0.0
Cross Entropy
1.5
Loopy BP
1.5
Swendsen?Wang
B5
BLib
Move 20
B5
BLib
Move 80
B5
BLib
Move 150
B5
BLib
Move 20
B5
BLib
Move 80
B5
BLib
Move 150
PN
Figure 4: Cross entropies N1 n [s0n log sn + (1 ? s0n ) log(1 ? sn )] between actual and
predicted territory outcomes, s0n and n for 327 Go positions. Sampling is compared
with Loopy BP (training and testing). 3 board positions were analysed for each
game (moves 20, 80 and 150). The Boltzmann5 (B5) and the BoltzmannLiberties
(BLib) models are compared.
Acknowledgements We thank I. Murray for helpful discussions on sampling and
T. Minka for general advice about probabilistic inference. This work was supported
by a grant from Microsoft Research UK.
References
[1] Thore Graepel, Mike Goutrie, Marco Kruger, and Ralf Herbrich. Learning on graphs
in the game of Go. In Proceedings of the International Conference on Artificial Neural
Networks, ICANN 2001, 2001.
[2] David Fotland.
Knowledge representation in the many faces of go.
ftp://www.joy.ne.jp/welcome/igs/Go/computer/mfg.tex.Z, 1993.
URL:
[3] Nicol N. Schrauldolph, Peter Dayan, and Terrance J. Sejnowski. Temporal difference
learning of position evaluation in the game of go. In Advances in Neural Information
Processing Systems 6, pages 817?824, San Fransisco, 1994. Morgan Kaufmann.
[4] Markus Enzenberger. The integration of a priori knowledge into a Go playing neural
network. URL: http://www.markus-enzenberger.de/neurogo.html, 1996.
[5] John Lafferty, Andrew McCallum, and Fernando Pereira. Conditional random fields:
Probabilistic models for segmenting and labeling sequence data. In Proc. Int. Conf.
on Machine Learning, 2001.
[6] Fabio Gagliardi Cozman. Generalizing variable elimination in Bayesian networks. In
Proceedings of the IBERAMIA/SBIA 2000 Workshops, pages 27?32, 2000.
[7] R. H. Swendsen and J-S Wang. Nonuniversal critical dynamics in Monte Carlo simulations. Physical Review Letters, 58:86?88, 1987.
[8] Robert G. Edwards and Alan D. Sokal. Generalisation of the Fortuin-KasteleynSwendsen-Wang representation and Monte Carlo algorithm. Physical Review Letters,
38(6), 1988.
[9] Yair Weiss. Belief propagation and revision in networks with loops. Technical report,
AI Lab Memo, MIT, Cambridge, 1998.
[10] A. L. Zobrist. Feature Extractions and Representations for Pattern Recognition and
the Game of Go. PhD thesis, Graduate School of the University of Wisconsin, 1970.
| 2688 |@word version:1 eliminating:1 c0:2 simulation:1 thoreg:1 score:3 gagliardi:1 interestingly:1 current:3 com:1 comparing:4 surprising:1 analysed:1 si:17 readily:1 john:1 update:1 joy:1 s0n:3 mccallum:1 affair:1 record:1 normalising:1 node:17 location:1 herbrich:1 org:2 firstly:1 evaluator:1 five:1 become:2 qualitative:1 incorrect:1 pairwise:1 expected:1 sbia:1 rapid:2 multi:1 actual:1 tex:1 increasing:2 revision:1 interpreted:2 temporal:2 every:1 uk:4 grant:1 segmenting:1 positive:2 safety:1 generalised:1 vertically:1 local:1 consequence:1 despite:1 black:25 exert:1 china:1 examined:1 co:1 limited:1 range:1 graduate:1 testing:1 recursive:1 aji:1 area:1 word:1 confidence:1 refers:1 suggest:1 convenience:1 close:1 selection:2 influence:5 www:5 go:44 starting:1 simplicity:1 assigns:1 rule:4 importantly:1 his:1 ralf:1 updated:1 play:1 exact:6 neighbouring:3 us:1 element:3 recognition:2 predicts:1 database:2 ising:2 bottom:2 mike:1 wang:11 capture:1 calculate:1 ferromagnetic:1 connected:1 intuition:1 complexity:5 cam:2 dynamic:1 trained:5 depend:1 segment:2 completely:1 cozman:1 surrounding:1 forced:1 describe:1 sejnowski:2 monte:2 artificial:1 labeling:1 outcome:11 choosing:1 whose:2 solve:1 valued:2 analyse:1 itself:1 final:4 ip:1 zobrist:1 sequence:1 propose:1 loop:1 alleviates:1 mixing:1 demon:1 empty:11 cluster:3 produce:3 ftp:1 coupling:7 andrew:1 ac:2 measured:2 ij:3 school:1 progress:1 edward:1 strong:1 predicted:3 indicate:1 lingers:2 direction:2 liberty:8 safe:1 programmer:1 elimination:2 decompose:1 mxy:2 secondly:1 boltzmann5:11 marco:1 around:1 considered:3 swendsen:10 exp:4 predict:1 bj:5 automate:1 proc:1 bond:8 knowl:1 create:2 successfully:1 occupying:1 reflects:1 mit:1 occupied:1 pn:1 structuring:1 boltzmannliberties:7 modelling:4 indicates:1 likelihood:4 sense:2 helpful:1 inference:10 dayan:2 sokal:1 typically:3 lj:1 interested:1 html:2 priori:1 spatial:3 integration:1 mackay:2 marginal:1 field:7 construct:1 equal:1 having:1 extraction:1 sampling:10 identical:1 future:2 others:1 report:1 serious:1 few:1 neighbour:3 winter:2 replaced:1 microsoft:3 n1:1 attempt:2 message:3 evaluation:10 chain:12 accurate:2 edge:3 necessary:1 xy:2 amateur:1 literally:1 tree:3 old:2 ancient:1 desired:1 plotted:1 circle:1 formalise:1 contiguous:1 loopy:10 ordinary:1 cost:1 vertex:11 subset:1 neutral:1 hundred:1 comprised:1 too:1 fransisco:1 damp:1 learnt:2 combined:1 confident:1 international:1 probabilistic:4 physic:2 together:1 connecting:1 connectivity:2 again:1 thesis:1 containing:1 slowly:1 possibly:1 dead:2 external:3 expert:5 conf:1 li:2 account:1 potential:5 de:1 int:1 explicitly:1 depends:2 performed:2 view:1 later:1 lab:1 square:3 spin:5 accuracy:2 kaufmann:1 who:1 likewise:1 sy:4 identify:1 weak:2 bayesian:3 territory:32 produced:3 carlo:2 ago:1 energy:1 involved:2 minka:1 associated:3 couple:1 dataset:1 knowledge:2 emerges:1 graepel:3 cj:10 appears:3 improved:4 april:1 wei:1 evaluated:1 marginalising:1 just:1 stage:1 correlation:4 until:1 hand:1 horizontal:1 assessment:1 propagation:4 thore:2 effect:2 true:1 hence:1 symmetric:1 death:1 white:20 adjacent:2 mnew:7 game:27 during:2 branching:1 stone:23 crf:1 performs:2 temperature:1 common:4 physical:2 winner:1 attached:1 defeat:1 jp:1 marginals:2 refer:3 cambridge:4 gibbs:5 ai:2 grid:2 similarly:1 moving:1 success:1 life:1 scoring:1 captured:2 morgan:1 additional:1 fernando:1 mix:1 mold:3 alan:1 technical:1 defeated:1 schraudolph:1 long:1 cross:4 qi:1 prediction:16 expectation:1 df:2 iteration:1 represent:2 tailored:1 adopting:1 achieved:1 diagram:2 unlike:1 probably:1 undirected:2 fate:5 incorporates:2 lafferty:1 extracting:1 vital:1 gave:1 nonuniversal:1 topology:3 opposite:1 competing:1 reduce:1 idea:1 cn:1 qj:1 b5:7 utility:1 colour:2 url:4 suffer:1 peter:1 useful:4 detailed:1 kruger:1 amount:1 welcome:1 mrao:1 simplest:1 http:3 occupy:1 estimated:1 arising:1 per:1 group:5 four:1 threshold:1 vast:2 graph:4 year:1 sum:5 inverse:1 letter:2 uncertainty:6 place:2 laid:1 reasonable:1 oscillation:1 draw:1 fortuin:1 layer:1 gnu:1 strength:1 bp:8 software:1 markus:2 argument:1 relatively:1 department:2 according:1 smaller:1 chess:1 invariant:1 taken:1 ln:1 turn:3 count:3 loose:1 mfg:3 end:9 reversal:2 opponent:2 apply:3 observe:1 alternative:1 yair:1 professional:3 original:2 top:1 running:1 cf:5 completed:1 graphical:1 exploit:1 chinese:1 murray:1 move:19 flipping:1 diagonal:1 gradient:1 win:1 fabio:1 thank:1 nx:1 reason:3 maximising:1 pointwise:1 relationship:2 difficult:2 robert:1 expense:1 negative:2 memo:1 stern:1 boltzmann:2 unknown:1 perform:4 allowing:1 vertical:1 markov:2 descent:1 anti:1 arbitrary:1 david:3 pair:2 specified:1 connection:2 usually:1 pattern:2 below:1 challenge:1 confidently:1 program:4 belief:7 critical:2 rely:1 force:2 predicting:3 minimax:1 representing:3 ne:1 sn:3 review:2 taste:2 acknowledgement:1 nicol:1 wisconsin:1 expect:1 interesting:1 degree:1 playing:2 surrounded:1 course:1 repeat:1 supported:1 bias:2 allow:1 perceptron:1 face:4 depth:1 evaluating:1 made:2 san:1 ig:1 sj:22 approximate:1 status:1 confirm:1 factorize:1 fotland:1 search:1 enzenberger:3 symmetry:2 expansion:1 japanese:2 icann:1 main:1 spread:1 komi:2 advice:1 board:27 ny:1 position:25 originated:1 wish:1 pereira:1 winning:2 sf:5 learns:1 down:1 derives:1 intrinsic:1 exists:1 workshop:1 adding:1 ci:10 phd:1 entropy:4 intersection:3 generalizing:1 likely:1 forming:1 visual:1 prevents:1 horizontally:1 corresponds:1 determines:1 conditional:6 included:1 determined:3 generalisation:2 called:1 experimental:1 player:9 support:1 arises:1 evaluate:2 |
1,859 | 2,689 | Modeling Nonlinear Dependencies in
Natural Images using Mixture of
Laplacian Distribution
Hyun Jin Park and Te Won Lee
Institute for Neural Computation, UCSD
9500 Gilman Drive, La Jolla, CA 92093-0523
{hjinpark, tewon}@ucsd.edu
Abstract
Capturing dependencies in images in an unsupervised manner is
important for many image processing applications. We propose a
new method for capturing nonlinear dependencies in images of
natural scenes. This method is an extension of the linear Independent
Component Analysis (ICA) method by building a hierarchical model
based on ICA and mixture of Laplacian distribution. The model
parameters are learned via an EM algorithm and it can accurately
capture variance correlation and other high order structures in a
simple manner. We visualize the learned variance structure and
demonstrate applications to image segmentation and denoising.
1
In trod u ction
Unsupervised learning has become an important tool for understanding biological
information processing and building intelligent signal processing methods. Real
biological systems however are much more robust and flexible than current artificial
intelligence mostly due to a much more efficient representations used in biological
systems. Therefore, unsupervised learning algorithms that capture more sophisticated
representations can provide a better understanding of neural information processing
and also provide improved algorithm for signal processing applications. For example,
independent component analysis (ICA) can learn representations similar to simple cell
receptive fields in visual cortex [1] and is also applied for feature extraction, image
segmentation and denoising [2,3]. ICA can approximate statistics of natural image
patches by Eq.(1,2), where X is the data and u is a source signal whose distribution is
a product of sparse distributions like a generalized Laplacian distribution.
X = Au
(1)
P (u ) = ? P (u i )
(2)
But the representation learned by the ICA algorithm is relatively low-level. In
biological systems there are more high-level representations such as contours,
textures and objects, which are not well represented by the linear ICA model. ICA
learns only linear dependency between pixels by finding strongly correlated linear
axis. Therefore, the modeling capability of ICA is quite limited. Previous approaches
showed that one can learn more sophisticated high-level representations by capturing
nonlinear dependencies in a post-processing step after the ICA step [4,5,6,7,8].
The focus of these efforts has centered on variance correlation in natural images. After
ICA, a source signal is not linearly predictable from others. However, given variance
dependencies, a source signal is still ?predictable? in a nonlinear manner. It is not
possible to de-correlate this variance dependency using a linear transformation.
Several researchers have proposed extensions to capture the nonlinear dependencies.
Portilla et al. used Gaussian Scale Mixture (GSM) to model variance dependency in
wavelet domain. This model can learn variance correlation in source prior and showed
improvement in image denoising [4]. But in this model, dependency is defined only
between a subset of wavelet coefficients. Hyvarinen and Hoyer suggested using a
special variance related distribution to model the variance correlated source prior.
This model can learn grouping of dependent sources (Subspace ICA) or topographic
arrangements of correlated sources (Topographic ICA) [5,6]. Similarly, Welling et al.
suggested a product of expert model where each expert represents a variance
correlated group [7]. The product form of the model enables applications to image
denoising. But these models don?t reveal higher-order structures explicitly.
Our model is motivated by Lewicki and Karklin who proposed a 2-stage model where
the 1st stage is an ICA model (Eq. (3)) and the 2 nd-stage is a linear generative model
where another source v generates logarithmic variance for the 1st stage (Eq. (4)) [8].
This model captures variance dependency structure explicitly, but treating variance as
an additional random variable introduces another level of complexity and requires
several approximations. Thus, it is difficult to obtain a simple analytic PDF of source
signal u and to apply the model for image processing problems.
(
P (u | ? ) = c exp ? u / ?
q
)
(3)
log[? ] = Bv
(4)
We propose a hierarchical model based on ICA and a mixture of Laplacian
distribution. Our model can be considered as a simplification of model in [8] by
constraining v to be 0/1 random vector where only one element can be 1. Our model is
computationally simpler but still can capture variance dependency. Experiments show
that our model can reveal higher order structures similar to [8]. In addition, our model
provides a simple parametric PDF of variance correlated priors, which is an important
advantage for adaptive signal processing. Utilizing this, we demonstrate simple
applications on image segmentation and image denoising. Our model provides an
improved statistic model for natural images and can be used for other applications
including feature extraction, image coding, or learning even higher order structures.
2
Modeling nonlinear dependencies
We propose a hierarchical or 2-stage model where the 1 st stage is an ICA source signal
model and the 2nd stage is modeled by a mixture model with different variances (figure
1). In natural images, the correlation of variance reflects different types of regularities
in the real world. Such specialized regularities can be summarized as ?context?
information. To model the context dependent variance correlation, we use mixture
models where Laplacian distributions with different variance represent different
contexts. For each image patch, a context variable Z ?selects? which Laplacian
distribution will represent ICA source signal u. Laplacian distributions have 0-mean
but different variances. The advantage of Laplacian distribution for modeling context
is that we can model a sparse distribution using only one Laplacian distribution. But
we need more than two Gaussian distributions to do the same thing. Also conventional
ICA is a special case of our model with one Laplacian. We define the mixture model
and its learning algorithm in the next sections.
Figure 1: Proposed hierarchical model (1st stage is ICA generative model. 2nd stage is
mixture of ?context dependent? Laplacian distributions which model U. Z is a random
variable that selects a Laplacian distribution that generates the given image patch)
2.1
Mixture of Laplacian Distribution
We define a PDF for mixture of M-dimensional Laplacian Distribution as Eq.(5),
where N is the number of data samples, and K is the number of mixtures.
N
N K
M
N K
r
r r
P(U | ?, ?) = ? P(u n | ?, ?) = ?? ? k P(u n | ?k ) = ?? ? k ?
n
n
k
n
k
m
1
(2? )
k ,m
? u n,m
exp? ?
? ?k , m
?
?
?
?
?
(5)
r
r r
r r
un = (un,1 , un , 2 , , , un,M ) : n-th data sample, U = (u1 , u 2 , , , ui , , , u N )
r
r r
r
r
?k = (?k ,1 , ?k , 2 ,..., ?k ,M ) : Variance of k-th Laplacian distribution, ? = (?1 , ?2 , , , ?k , , , ?K )
?k
: probability of Laplacian distribution k, ? = (? 1 , , , ? K ) and
?
k
?k =1
It is not easy to maximize Eq.(5) directly, and we use EM (expectation maximization)
algorithm for parameter estimation. Here we introduce a new hidden context variable
Z that represents which Laplacian k, is responsible for a given data point. Assuming
we know the hidden variable Z, we can write the likelihood of data and Z as Eq.(6),
zkn
?K ?
??
N
?
u n ,m
r
? ? 1 ??
z kn
?
?
P(U , Z | ?, ? ) = ? P(u n , Z | ?, ? ) = ? ? (? k ) ? ? ?
? exp? ? z kn
?
?
?
?
?k , m
n
n
k
m ? ? 2?k ,m ?
?
?? ??
?
N
? ??? ??
? ?
? ??? ?
? ??? ??
(6)
z kn : Hidden binary random variable, 1 if n-th data sample is generated from k-th
Laplacian, 0 other wise. ( Z = (z kn ) and ? z kn = 1 for all n = 1?N)
k
2.2
EM algorithm for learning the mixture model
The EM algorithm maximizes the log likelihood of data averaged over hidden variable
Z. The log likelihood and its expectation can be computed as in Eq.(7,8).
?
?
u
1
log P(U , Z | ?, ? ) = ? ? z kn log(? k ) + ? z kn ? log(
) ? n ,m
?
2
?
?
n ,k ?
m
k ,m
k ,m
?
?
??
??
??
??
(7)
?
?
u
1
E {log P (U , Z | ?, ? )} = ? E z kn ?log(? k ) + ? ? log(
) ? n ,m
?
2? k , m
?k , m
n ,k
m ?
??
{ }
??
??
??
??
(8)
The expectation in Eq.(8) can be evaluated, if we are given the data U and estimated
parameters ? and ?. For ? and ?, EM algorithm uses current estimation ?? and ??.
{ } {
} ? z P( z
E z kn ? E z kn | U , ? ' , ? ' =
1
zkn =0
n
k
n
k
| u n , ?' , ? ' ) = P ( z kn = 1 | u n , ? ' , ? ' )
(9)
=
P (u n | z kn = 1, ?' , ? ' ) P( z kn = 1 | ? ' , ? ' )
P(u n | ?' , ? ' )
=
M
u n ,m
1
1
1
exp(?
) ?? k ' =
?
?k , m '
P (u n | ? ' , ? ' ) m 2?k ,m '
cn
M
?k '
? 2?
m
k ,m
'
exp(?
u n ,m
?k , m '
)
Where the normalization constant can be computed as
K
K
M
k
k =1
m =1
cn = P (u n | ? ' , ? ' ) = ? P (u n | z kn , ? ' , ? ' ) P ( z kn | ? ' , ? ' ) = ? ? k ?
1
(2? )
exp( ?
k ,m
u n ,m
?k ,m
)
(10)
The EM algorithm works by maximizing Eq.(8), given the expectation computed from
Eq.(9,10). Eq.(9,10) can be computed using ?? and ?? estimated in the previous
iteration of EM algorithm. This is E-step of EM algorithm. Then in M-step of EM
algorithm, we need to maximize Eq.(8) over parameter ? and ?.
First, we can maximize Eq.(8) with respect to ?, by setting the derivative as 0.
?? 1
u n,m ??
?E{log P (U , Z | ?, ? )}
?? = 0
= ? E z kn ?? ?
+
?
?? k ,m
?
(? k ,m ) 2 ???
???
n
k ,m
?
{ }
? ? k ,m
? E{z }? u
=
? E{z }
n
k
n ,m
n
(11)
n
k
n
Second, for maximization of Eq.(8) with respect to ?, we can rewrite Eq.(8) as below.
(12)
E {log P (U , Z | ? , ? )} = C + ? E {z kn' }log(? k ' )
n ,k '
As we see, the derivative of Eq.(12) with respect to ? cannot be 0. Instead, we need to
use Lagrange multiplier method for maximization. A Lagrange function can be
defined as Eq.(14) where ? is a Lagrange multiplier.
{ }
(13)
L (? , ? ) = ? ? E z kn' log(? k ' ) + ? (? ? k ' ? 1)
n,k '
k'
By setting the derivative of Eq.(13) to be 0 with respect to ? and ?, we can simply get
the maximization solution with respect to ?. We just show the solution in Eq.(14).
?L(?, ? )
?L(?, ? )
=0
= 0,
??
??
?
? ?
?
? ? k = ? ? E z kn ? / ? ?? E z kn ?
?
? ? k n
? n
{ }
{ }
(14)
Then the EM algorithm can be summarized as figure 2. For the convergence criteria,
we can use the expectation of log likelihood, which can be calculated from Eq. (8).
?k =
{ }
, ?k , m = E um + e (e is small random noise)
2. Calculate the Expectation by
1. Initialize
1
K
u n ,m
1 M ?k '
exp( ?
)
?
c n m 2? k , m '
? k ,m '
3. Maximize the log likelihood given the Expectation
{ } {
}
E z kn ? E z kn | U , ? ' , ? ' =
?
? ?
?
?k ,m ? ? ? E {z kn }? u n,m ? / ? ? E {z kn }? ,
?
? ?
?
? k ? ? ? E {z kn }? / ? ?? E {z kn }?
? ? n
?
? n
? ? k n
?
4. If (converged) stop, otherwise repeat from step 2.
?
n
Figure 2: Outline of EM algorithm for Learning the Mixture Model
3
Experimental Results
Here we provide examples of image data and show how the learning procedure is
performed for the mixture model. We also provide visualization of learned variances
that reveal the structure of variance correlation and an application to image denoising.
3.1
Learning Nonlinear Dependencies in Natural images
As shown in figure 1, the 1 st stage of the proposed model is simply the linear ICA. The
ICA matrix A and W(=A-1) are learned by the FastICA algorithm [9]. We sampled
105(=N) data from 16x16 patches (256 dim.) of natural images and use them for both
first and second stage learning. ICA input dimension is 256, and source dimension is
set to be 160(=M). The learned ICA basis is partially shown in figure 1. The 2nd stage
mixture model is learned given the ICA source signals. In the 2 nd stage the number of
mixtures is set to 16, 64, or 256(=K). Training by the EM algorithm is fast and several
hundred iterations are sufficient for convergence (0.5 hour on a 1.7GHz Pentium PC).
For the visualization of learned variance, we adapted the visualization method from
[8]. Each dimension of ICA source signal corresponds to an ICA basis (columns of A)
and each ICA basis is localized in both image and frequency space. Then for each
Laplacian distribution, we can display its variance vector as a set of points in image
and frequency space. Each point can be color coded by variance value as figure 3.
(a1)
(a2)
(b1)
(b2)
Figure 3: Visualization of learned variances (a1 and a2 visualize variance of
Laplacian #4 and b1 and 2 show that of Laplacian #5. High variance value is mapped
to red color and low variance is mapped to blue. In Laplacian #4, variances for
diagonally oriented edges are high. But in Laplacian #5, variances for edges at
spatially right position are high. Variance structures are related to ?contexts? in the
image. For example, Laplacian #4 explains image patches that have oriented textures
or edges. Laplacian #5 captures patches where left side of the patch is clean but right
side is filled with randomly oriented edges.)
A key idea of our model is that we can mix up independent distributions to get nonlinearly dependent distribution. This modeling power can be shown by figure 4.
Figure 4: Joint distribution of nonlinearly dependent sources. ((a) is a joint histogram
of 2 ICA sources, (b) is computed from learned mixture model, and (c) is from learned
Laplacian model. In (a), variance of u2 is smaller than u1 at center area (arrow A), but
almost equal to u1 at outside (arrow B). So the variance of u2 is dependent on u1. This
nonlinear dependency is closely approximated by mixture model in (b), but not in (c).)
3.2
Unsupervised Image Segmentation
The idea behind our model is that the image can be modeled as mixture of different
variance correlated ?contexts?. We show how the learned model can be used to
classify different context by an unsupervised image segmentation task. Given learned
model and data, we can compute the expectation of a hidden variable Z from Eq. (9).
Then for an image patch, we can select a Laplacian distribution with highest
probability, which is the most explaining Laplacian or ?context?. For segmentation,
we use the model with 16 Laplacians. This enables abstract partitioning of images and
we can visualize organization of images more clearly (figure 5).
Figure 5: Unsupervised image segmentation (left is original image, middle is color
labeled image, right image shows color coded Laplacians with variance structure.
Each color corresponds to a Laplacian distribution, which represents surface or
textural organization of underlying contexts. Laplacian #14 captures smooth surface
and Laplacian #9 captures contrast between clear sky and textured ground scenes.)
3.3
Application to Image Restoration
The proposed mixture model provides a better parametric model of the ICA source
distribution and hence an improved model of the image structure. An advantage is in
the MAP (maximum a posterior) estimation of a noisy image. If we assume Gaussian
noise n, the image generation model can be written as Eq.(15). Then, we can compute
MAP estimation of ICA source signal u by Eq.(16) and reconstruct the original image.
(15)
X = Au + n
(16)
u? = argmax log P (u | X , A) = argmax (log P ( X | u , A) + log P (u ) )
u
u
Since we assumed Gaussian noise, P(X|u,A) in Eq. (16) is Gaussian. P(u) in Eq. (16)
can be modeled as a Laplacian or a mixture of Laplacian distribution. The mixture
distribution can be approximated by a maximum explaining Laplacian. We evaluated
3 different methods for image restoration including ICA MAP estimation with simple
Laplacian prior, same with Laplacian mixture prior, and the Wiener filter. Figure 6
shows an example and figure 7 summarizes the results obtained with different noise
levels. As shown MAP estimation with the mixture prior performs better than the
others in terms of SNR and SSIM (Structural Similarity Measure) [10].
Figure 6: Image restoration results (signal variance 1.0, noise variance 0.81)
16
ICA MAP (Mixture prior)
ICA MAP (Laplacian prior)
W iener
14
0.8
SSIM Index
SNR
12
10
8
6
0.6
0.4
0.2
4
2
ICA MAP(Mixture prior)
ICA MAP(Laplacian prior)
W iener
Noisy Image
1
0
0.5
1
1.5
Noise variance
2
2.5
0
0
0.5
1
1.5
Noise variance
2
2.5
Figure 7: SNR and SSIM for 3 different algorithms (signal variance = 1.0)
4
D i s c u s s i on
We proposed a mixture model to learn nonlinear dependencies of ICA source signals
for natural images. The proposed mixture of Laplacian distribution model is a
generalization of the conventional independent source priors and can model variance
dependency given natural image signals. Experiments show that the proposed model
can learn the variance correlated signals grouped as different mixtures and learn highlevel structures, which are highly correlated with the underlying physical properties
captured in the image. Our model provides an analytic prior of nearly independent and
variance-correlated signals, which was not viable in previous models [4,5,6,7,8].
The learned variances of the mixture model show structured localization in image and
frequency space, which are similar to the result in [8]. Since the model is given no
information about the spatial location or frequency of the source signals, we can
assume that the dependency captured by the mixture model reveals regularity in the
natural images. As shown in image labeling experiments, such regularities correspond
to specific surface types (textures) or boundaries between surfaces.
The learned mixture model can be used to discover hidden contexts that generated
such regularity or correlated signal groups. Experiments also show that the labeling of
image patches is highly correlated with the object surface types shown in the image.
The segmentation results show regularity across image space and strong correlation
with high-level concepts.
Finally, we showed applications of the model for image restoration. We compare the
performance with the conventional ICA MAP estimation and Wiener filter. Our
results suggest that the proposed model outperforms other traditional methods. It is
due to the estimation of the correlated variance structure, which provides an improved
prior that has not been considered in other methods.
In our future work, we plan to exploit the regularity of the image segmentation result
to lean more high-level structures by building additional hierarchies on the current
model. Furthermore, the application to image coding seems promising.
References
[1] A. J. Bell and T. J. Sejnowski, The ?Independent Components? of Natural Scenes are Edge
Filters, Vision Research, 37(23):3327?3338, 1997.
[2] A. Hyvarinen, Sparse Code Shrinkage: Denoising of Nongaussian Data by Maximum
Likelihood Estimation,Neural Computation, 11(7):1739-1768, 1999.
[3] T. Lee, M. Lewicki, and T. Sejnowski., ICA Mixture Models for unsupervised
Classification of non-gaussian classes and automatic context switching in blind separation.
PAMI, 22(10), October 2000.
[4] J. Portilla, V. Strela, M. J. Wainwright and E. P Simoncelli, Image Denoising using Scale
Mixtures of Gaussians in the Wavelet Domain, IEEE Trans. On Image Processing, Vol.12, No.
11, 1338-1351, 2003.
[5] A. Hyvarinen, P. O. Hoyer. Emergence of phase and shift invariant features by
decomposition of natural images into independent feature subspaces. Neurocomputing, 1999.
[6] A. Hyvarinen, P.O. Hoyer, Topographic Independent component analysis as a model of V1
Receptive Fields, Neurocomputing, Vol. 38-40, June 2001.
[7] M. Welling and G. E. Hinton, S. Osindero, Learning Sparse Topographic Representations
with Products of Student-t Distributions, NIPS, 2002.
[8] M. S. Lewicki and Y. Karklin, Learning higher-order structures in natural images, Network:
Comput. Neural Syst. 14 (August 2003) 483-499.
[9] A.Hyvarinen, P.O. Hoyer, Fast ICA matlab code.,
http://www.cis.hut.fi/projects/compneuro/extensions.html/
[10] Z. Wang, A. C. Bovik, H. R. Sheikh and E. P. Simoncelli, The SSIM Index for Image
Quality Assessment, IEEE Transactions on Image Processing, vol. 13, no. 4, Apr. 2004.
| 2689 |@word middle:1 seems:1 nd:5 decomposition:1 outperforms:1 current:3 written:1 enables:2 analytic:2 treating:1 intelligence:1 generative:2 provides:5 location:1 simpler:1 become:1 viable:1 manner:3 introduce:1 ica:38 project:1 discover:1 underlying:2 maximizes:1 strela:1 finding:1 transformation:1 sky:1 um:1 partitioning:1 textural:1 switching:1 pami:1 au:2 limited:1 averaged:1 responsible:1 procedure:1 area:1 bell:1 suggest:1 get:2 cannot:1 context:14 www:1 conventional:3 map:9 center:1 maximizing:1 utilizing:1 hierarchy:1 us:1 element:1 gilman:1 approximated:2 lean:1 labeled:1 wang:1 capture:8 calculate:1 highest:1 predictable:2 complexity:1 ui:1 rewrite:1 localization:1 basis:3 textured:1 joint:2 represented:1 fast:2 ction:1 sejnowski:2 artificial:1 labeling:2 outside:1 whose:1 quite:1 otherwise:1 reconstruct:1 statistic:2 topographic:4 emergence:1 noisy:2 advantage:3 highlevel:1 propose:3 product:4 convergence:2 regularity:7 object:2 eq:25 strong:1 closely:1 filter:3 centered:1 explains:1 generalization:1 biological:4 extension:3 hut:1 considered:2 ground:1 exp:7 visualize:3 a2:2 estimation:9 grouped:1 tool:1 reflects:1 clearly:1 gaussian:6 shrinkage:1 focus:1 june:1 improvement:1 likelihood:6 pentium:1 contrast:1 dim:1 dependent:6 hidden:6 selects:2 pixel:1 classification:1 flexible:1 html:1 plan:1 spatial:1 special:2 initialize:1 field:2 equal:1 extraction:2 represents:3 park:1 unsupervised:7 nearly:1 future:1 others:2 intelligent:1 oriented:3 randomly:1 neurocomputing:2 argmax:2 phase:1 organization:2 highly:2 introduces:1 mixture:34 pc:1 behind:1 edge:5 trod:1 filled:1 column:1 modeling:5 classify:1 restoration:4 maximization:4 subset:1 snr:3 hundred:1 fastica:1 osindero:1 dependency:18 kn:26 st:5 lee:2 nongaussian:1 expert:2 derivative:3 syst:1 de:1 coding:2 summarized:2 b2:1 coefficient:1 student:1 explicitly:2 blind:1 performed:1 red:1 capability:1 wiener:2 variance:47 who:1 correspond:1 accurately:1 drive:1 researcher:1 converged:1 gsm:1 frequency:4 stop:1 sampled:1 color:5 segmentation:9 sophisticated:2 higher:4 improved:4 evaluated:2 strongly:1 furthermore:1 just:1 stage:13 correlation:7 nonlinear:9 assessment:1 quality:1 reveal:3 building:3 concept:1 multiplier:2 hence:1 spatially:1 won:1 criterion:1 generalized:1 pdf:3 outline:1 demonstrate:2 performs:1 image:62 wise:1 fi:1 specialized:1 physical:1 automatic:1 similarly:1 cortex:1 surface:5 similarity:1 posterior:1 showed:3 jolla:1 binary:1 captured:2 additional:2 maximize:4 signal:20 mix:1 simoncelli:2 smooth:1 post:1 coded:2 a1:2 laplacian:39 vision:1 expectation:8 iteration:2 represent:2 normalization:1 histogram:1 cell:1 addition:1 source:21 bovik:1 thing:1 structural:1 constraining:1 easy:1 idea:2 cn:2 shift:1 motivated:1 effort:1 zkn:2 matlab:1 clear:1 http:1 estimated:2 blue:1 write:1 vol:3 group:2 key:1 clean:1 v1:1 almost:1 patch:9 separation:1 summarizes:1 capturing:3 simplification:1 display:1 bv:1 adapted:1 scene:3 generates:2 u1:4 relatively:1 structured:1 smaller:1 across:1 em:12 sheikh:1 invariant:1 computationally:1 visualization:4 know:1 gaussians:1 apply:1 hierarchical:4 original:2 exploit:1 arrangement:1 receptive:2 parametric:2 traditional:1 hoyer:4 subspace:2 mapped:2 assuming:1 code:2 modeled:3 index:2 difficult:1 mostly:1 october:1 ssim:4 hyun:1 jin:1 hinton:1 portilla:2 ucsd:2 august:1 nonlinearly:2 learned:15 hour:1 nip:1 trans:1 suggested:2 below:1 laplacians:2 including:2 wainwright:1 power:1 natural:14 karklin:2 tewon:1 axis:1 prior:13 understanding:2 generation:1 localized:1 sufficient:1 diagonally:1 repeat:1 side:2 institute:1 explaining:2 sparse:4 ghz:1 boundary:1 calculated:1 dimension:3 world:1 contour:1 adaptive:1 hyvarinen:5 correlate:1 welling:2 transaction:1 approximate:1 reveals:1 b1:2 assumed:1 don:1 un:4 promising:1 learn:7 robust:1 ca:1 domain:2 apr:1 linearly:1 arrow:2 noise:7 x16:1 position:1 comput:1 learns:1 wavelet:3 specific:1 grouping:1 ci:1 texture:3 te:1 logarithmic:1 simply:2 visual:1 lagrange:3 partially:1 lewicki:3 u2:2 corresponds:2 denoising:8 experimental:1 la:1 select:1 correlated:12 |
1,860 | 269 | Predicting Weather Using a Genetic Memory
Predicting Weather Using a Genetic Memory: a Combination of Kanerva's Sparse Distributed Memory with
Holland's Genetic Algorithms
David Rogers
Research Institute for Advanced Computer Science
MS 230-5, NASA Ames Research Center
Moffett Field, CA 94035
ABSTRACT
Kanerva's sparse distributed memory (SDM) is an associative-memory model based on the mathematical properties of high-dimensional
binary address spaces. Holland's genetic algorithms are a search technique for high-dimensional spaces inspired by evolutionary processes
of DNA. "Genetic Memory" is a hybrid of the above two systems,
in which the memory uses a genetic algorithm to dynamically reconfigure its physical storage locations to reflect correlations between
the stored addresses and data. For example, when presented with
raw weather station data, the Genetic Memory discovers specific features in the weather data which correlate well with upcoming rain,
and reconfigures the memory to utilize this information effectively.
This architecture is designed to maximize the ability of the system
to scale-up to handle real-world problems.
INTRODUCTION
The future success of neural networks depends on an ability to "scale-up" from
small networks and low-dimensional toy problems to networks of thousands or millions of nodes and high-dimensional real-world problems. (The dimensionality of a
problem refers to the number of variables needed to describe the problem domain.)
Unless neural networks are shown to be scalable to real-world problems, they will
likely remain restricted to a few specialized applications.
Scaling-up adds two types of computational demands to a system. First, there is a
linear increase in computational demand proportional to the increased number of variables. Second, there is a greater, nonlinear increase in computational demand due to
455
456
Rogers
the number of interactions that can occur between the variables. This latter effect is
primarily responsible for the difficulties encountered in scaling-up many systems.
In general, it is difficult to scale-up a system unless it is specifically designed to
function well in high-dimensional domains.
Two systems designed to function well in high-dimensional domains are Kanerva' s
sparse distributed memory (Kanerva, 1988) and Holland's genetic algorithms
(Holland, 1986). I hypothesized that a hybrid of these two systems would preserve
this ability to operate well in high-dimensional environments, and offer grater functionality than either individually. I call this hybrid Genetic Memory. To test its
capabilities, I applied it to the problem of forecasting rain from local weather data.
Kanerva's sparse distributed memory (SDM) is an associative-memory model based
on the mathematical properties of high-dimensional binary address spaces. It can be
represented as a three-layer neural-network with an extremely large nwnber of
nodes (I ,000,000+) in the middle layer. In its standard formulation, the connections between the input layer and the hidden layer (the input representation used by
the system) are flXed, and learning is done by changing the values of the connections
between the hidden layer and the output layer.
Holland's genetic algorithms are a search technique for high-dimensional spaces inspired by evolutionary processes of DNA. Members of a set of binary strings competes for the opportunity to recombine. Recombination is done by selecting two
"successful" members of the population to be the parents. A new string is created
by splicing together pieces of each parent. Finally, the new string is placed into the
set, and some "unsuccessful" older string removed.
"Genetic Memory" is a hybrid of the above two systems. In this hybrid, a genetic algorithm is used to reconfigure the connections between the input layer and the hidden layer. The connections between the hidden layer and the output layer are
changed using the standard method for a sparse distributed memory. The "success"
of an input representation is determined by how well it reflects correlations between addresses and data, using my previously presented work on statistical prediction (Rogers, 1988). Thus, we have two separate learning algorithms in the two levels. The memory uses the genetic algorithm to dynamically reconfigure its input
representation to better reflect correlations between collections of input variables
and the stored data.
I applied this Genetic Memory architecture to the problem of predicting rain given
only local weather features such as the air pressure, the cloud cover, the month, the
temperature, etc. The weather data contained 15 features, sampled every 4-hours
over a 2O-year period on the Australian coast. I coded each state into a 256-bit address, and stored at that address a single bit which denoted whether it rained in the 4
hours following that weather state. I allowed the genetic algorithm to reconfigure
the memory while it scanned the file of weather states.
The success of this procedure was measured in two ways. First, once the training
was completed, the Genetic Memory was better at predicting rain than was the standard sparse distributed memory. Second, I had access to the input representations
discovered by the Genetic Memory and could view the specific combinations of features that predicted rain. Thus, unlike many neural networks, the Genetic Memory
allows the user to inspect the internal representations it discovers during training.
Predicting Weather Using a Genetic Memory
Reference Address
01010101101
~
~
Input Data
Radius
o
I 0 I 01
Dist
Select
1010101010
1011101100
0010101111
1101101101
0100000110
0110101001
1011010110
1100010111
""
? _;1
1101100111
0000011110
0011011001
I tI 0 11 I
+ + + ++ ,
0111 11 11 0
?
?
-1 -1
-1
1
1
II
1 -1
1 -1 1
?
..._--
?
O?
?
0 ?
?
?
0 ?
1?
-1 -1
-1
1
1
1 -1
1 -1
o
-2 0
2
2
0
0 0
-1
-1 -1
1 1
1
-1
1
0 0
1 -1 1
1 1 1 1 1 100 1 1
Location
Addresses
Data
Counters
, , , , + ++ ++
'L
I -31 31 -31 3 I
Threshold at 0 + + + + + + + + + +
Output Data I 0 I 0 I 0 I tI tI tI 0 I 1 I 0 11 I
Sums 1-31
-s 1-31 sis 13
Figure 1: Structure of a Sparse Distributed Memory
SPARSE DISTRmUTED MEMORY
Sparse distributed memory can be best illustrated as a variant of random-access memory (RAM). The structure of a twelve-location SDM with ten-bit addresses and
ten-bit data is shown in figure 1.
A memory location is a row in this figure. The location addresses are set to random
addresses. The data counters are initialized to zero. All operations begin with addressing the memory; this entails fmding the Hamming distance between the reference address and each of the location addresses. If this distance is less than or equal
to the Hamming radius. the select-vector entry is set to 1. and that location is
termed selected. The ensemble of such selected locations is called the selected set.
Selection is noted in the figure as non-gray rows. A radius is chosen so that only a
small percentage of the memory locations are selected for a given reference address.
When writing to the memory. all selected counters beneath elements of the input data equal to 1 are incremented. and all selected counters beneath elements of the input
data equal to 0 are decremented. This completes a write operation. When reading
from the memory. the selected data counters are summed columnwise into the register swns. If the value of a sum is greater than or equal to zero. we set the corresponding bit in the output data to 1; otherwise. we set the bit in the output data to
O. (When reading. the contents of the input data are ignored.)
457
458
Rogers
This example makes clear that a datum is distributed over the data counters of the selected locations when writing. and that the datum is reconstructed during reading by
averaging the sums of these counters. However, depending on what additional data
were written into some of the selected locations, and depending on how these data
correlate with the original data. the reconstruction may contain noise.
The SDM model can also be described as a fully-connected three-layer feed-forward
neural network. In this model. the location addresses are the weights between the
input layer and the hidden units. and the data counters are the weights between the
hidden units and the output layer. Note that the number of hidden-layer nodes (at
least 1,000 and possibly up to 1,000,(00) is much larger than is commonly used for
artificial neural networks. It is unclear how well standard algorithms. such as backpropagation, would perform with such a large number of units in the hidden layer.
HOLLAND'S GENETIC ALGORITHMS
Genetic Algorithms are a search technique for high-dimensional spaces inspired by
the evolutionary processes of DNA. The domain of a genetic algorithm is a population ofrued-length binary strings and a fitness function, which is a method for evaluating the fitness of each of the members. We use this fitness function to select two
highly-ranked members for recombination. and one lowly-ranked member for replacement (The selection may be done either absolutely. with the best and worst
members always being selected. or probabilisticly. with the members being chosen
proportional to their fitness scores.)
The member selected as bad is removed from the population. The two members selected as good are then recombined to create a new member to take its place in the
population. In effect, the genetic algorithm is a search over a high-dimensional
space for strings which are highly-rated by the fitness function.
The process used to create new members of the population is called crossover. In a
crossover. we align the two good candidates end-to-end and segment them at one or
more crossover-points. We then create a new string by starting the transcription of
bits at one of the parent strings, and switching the transcription to the other parent
at the crossover-points. This new string is placed into the population. taking the
place of the poorly-rated member.
11??I'~IOllllll ????
111:1"1.1 10 ...
Second parent
1101101101:
~
~
First parent
~
New member
Figure 2: Crossover of two binary strings
By running the genetic algorithm over the population many times, the population
"evolves" towards members which are rated more fit by our fitness function.
Predicting Weather Using a Genetic Memory
Output Layer
Weights changed using
perceptron rule
Hidden Unit Layer
-..
Weights changed using
Genetic Algorithm
Input Layer
Figure 3: Structure of a Genetic Memory
Holland has a lIIaUIt~lIIaUl:at pruuI Uta&. g~l1~Ul: atgunuul1s uastAI UI1 Ule crossover
procedure are an extremely efficient method for searching a high-dimensional space.
GENETIC MEMORY
Genetic Memory is a hybrid of Kanerva's Sparse Distributed Memory and Holland's
Genetic Algorithms. In this hybrid, the location addresses of the SDM are not held
constant; rather, a Genetic Algorithm is used to move them to more advantageous
positions in the address space. If we view SDM as a neural net, this hybrid uses a genetic algorithm to change the weights in the connections between the input layer and
the hidden unit layer, while the connections between the hidden unit layer and the
output layer at changed using the standard method for a SDM.
Most other work which combined neural networks and genetic algorithms kept multiple networks; the Genetic Algorithm was used to recombine the more successful
of these networks to create new entire networks.
In a Genetic Memory there is a single network with different algorithms changing
the weights in different layers. Thus, a Genetic Memory incorporates the Genetic
Algorithm directly into the operation of a single network.
AUSTRALIAN WEATHER DATA
The weather data was collected at a single site on the Australian coast. A sample
was taken every 4 hours for 25 years; the me contains over 58,000 weather samples
The file contained 15 distinct features, including year, month, day of the month,
time of day, pressure, dry bulb temperature, wet bulb temperature, dew point, wind
speed, wind direction, cloud cover, and whether it rained in the past four hours.
For this work, I coded each weather sample into a 256-bit word. Each weather sample was coded into a 256-bit binary address, giving each feature a 16-bit field in that
address. The feature values were coarse-coded into a simple thennometer-style
code. For example, figure 4 shows the code used for month.
PROCEDURE FOR WEATHER PREDICTION
In the standard SDM model, the locations addresses are held constant. In a Genetic
Memory, the location addresses are reconfigured using a Genetic Algorithm.
459
460
Rogers
JAN:
FEB:
MAR:
APR:
MAY:
JUN:
1111111100000000
0111111111000000
0011111111100000
0000111111110000
0000011111111000
0000001111111110
JUL: 1000000001111111
AUG:
SEP:
OCT:
NOV:
DEC:
1100000000111111
1111000000011111
1111100000001111
1111110000000011
1111111000000001
Figure 4: 16-bit code used for month
The fitness function used is based on my work on statistical prediction and presented
at NIPS-88 [Rogers 1988]. This work assigns a number to each physical storage location (a row in the figure) which is a measure of the predictiveness of that location.
Highly-predictive locations are recombined using crossover; the newly-created location address is given to a location which is relatively unpredictive. The data
counter is a measure of the co"elation between the selection of a location and the
occurrence of a given bit value. Thus, we can use the data counters to judge the fitness, i.e., the predictiveness, of each memory location.
To train the memory, we present the memory with each weather state in turn. The
memory is not shown the data a multiple number of times. For each state, the memory is addressed with the 256-bit address which represents it. non is written to the
memory if it does not rain in the next four hours, and "1" if it does. After the memory has seen a given number of weather samples, the Genetic Algorithm is performed to replace a poorly-predictive location with a new address created from two
predictive addresses.
The procedure is continued until the memory has seen 50,000 weather samples, and
has performed -5,000 genetic recombinations.
ANAL YSIS OF RESULTS
The initial results from the Genetic Memory procedure was conducted on a memory
with 1,000 storage locations. The weather sample set consisted of a sequence of
weather samples taken every 4 hours over a period of 20 years. In the sample set, it
rained in the next 4 hours for -10% of the samples, and was dry in the next four
hours in -90% of the samples.
The Genetic Memory was testing by storing -50,000 weather samples. The samples
were given to the memory in chronological order. During the course of storage, the
memory reconfigured itself with -5,000 genetic recombinations. A Genetic Memory and a standard Sparse Distributed Memory were tested against 1,000 previously
unseen weather samples. In initial experiments, the Genetic Memory had 50% fewer
errors than the Sparse Distributed Memory.
However, the Genetic Memory does not only show an improvement in performance,
it allows the user to analyze the genetically-determined memory locations to discover how the memory improved its performance.
By studying highly-rated memory locations in the Genetic Memory, we can open the
black box: that is, access the parameters the memory has decided are the most effective in associating the sample addresses with the sample data. This ability to access
the parameters the system found effective has two important implications. First,
Predicting Weather Using a Genetic Memory
the parameters may offer insights into the underlying physical processes in the system under study. Second. knowledge of how the system predicts may be vital for determining the robustness and the envelope of applicability of the memory prior to
embedding into a real-world system.
Simply scoring the performance of a system is not enough. We must be able to
"open the black box" to study why the system performs as it does.
OPENING THE BLACK BOX
When the training is completed. we can analyze the structure of memory locations
which performed well to discover which features they found most discriminatory
and which values of those features were preferred. For example. here is a memory
location which was rated highly-fit for predicting rain after training:
1101001100000011 1111011110101011 0111111100010000 1100000011011010
0100110011111011 1111110000000011 0111111011()()()()()() 001110110110()l10
000000101111011001100000010000100001001110110100 0100000111111111
0000000111111110 0000000011111111 0011011111111111 0100110000001000
By measuring the distance between a given 16-bit field and all possible values for
that field. we can discover which values of the feature are most desired. (Closer in
hamming distance is better.) The absolute range of values is the sensitivity of the location to changes along that feature dimension. Figure 5 shows an analysis of the 16bit field for month in the given memory location:
Location's 16-bit field
for month:0111111100010000
t
values for months
Distance
Less desirable
JAN: 1111111100000000
2
Value desirability
FEB: 0111111110000000
2
More desirable
MAR: 0011111111000000
4
APR: 0000111111110000
6
. .. etc ...
~
Feature (sensitivity
Month (12)
IOVf\ IS f-
S
f-
?
I
I
-
\-
11
JFMAMJJASOND
Values
Figure 5: Analyzing a location field
In this case. the location finds January and February the most desirable months for
rain. and July and August the least desirable months.
The relative sensitivity towards different features measures which features are most
important in making the prediction of rain. In this case. we have a change of distance of 12 bits. which makes this location very sensitive to the value of the month.
We can estimate which features are the most important in predicting rain by looking
at the relative sensitivity of the different fields in the location to changes in their
feature. The following graphs show the most sensitive features of the previously
shown memory location towards predicting rain; that is. the location is very sensitive to the combination of all these fields with the proper values.
461
462
Rogers
Dry bulb temp (12)
Cloud cover (13)
Iii
Iii
10
ou.........&.J...u...o....o.J....&............"Ju.........~
None
Low
oL........L.~...L..L..J..........L...............-"--'-J
210
High
Pressure (12)
270
240
Month (12)
IIi
Iii
10
10100
JFMAMJJASOND
Figure 6: The four most sensitive features
The "most preferred values" of these fields are the minima of these graphs. For example, this location greatly prefers January and February over June and JUly. The preferences of this location are for the month to be January or February, for low pressure. high cloud cover, and low temperature. Surprisingly. whether it rained in the
last four hours is not one of the most important features for this location.
We can also look some of the least sensitive features. The following graphs show
the least sensitive features of the memory location towards predicting rain; that is,
the location is relatively insensitive to the values of these features.
Wet bulb temp (5)
Year (5)
Wind direction (4)
Iii
10
61
73
80
210
240
270
N
E
S
W
Figure 7: The three least sensitive features
This set contains some fields that one would expect to be relatively unimportant,
such as year. Fields such as wind direction is unimportant to this location, but interestingly other highly-rated locations fmd it to be very useful in other regions of the
weather space.
Predicting Weather Using a Genetic Memory
COMPARISON WITH DAVIS' METHOD
Davis' Algorithm has been shown to be a powerful new method for augmenting the
power of a backpropagation-based system. The following is an attempt to contrast
our approaches, without denigrating the importance his groundbreaking work. The
reader is referred to his book for detailed information about his approach.
It is difficult to directly compare the performance of these techniques given the preliminary nature of the experiments done with Genetic Memory. However, it is possible to compare architectural features of the systems, and estimate the relative
strengths a weaknesses.
? 8ackpropagation vs. Associative Memories: Davis' approach relies on the performance of the backpropagation algorithm for the central learning cycle of the system. Associative memories have a far quicker learning cycle than backpropagation
networks, and have been shown to have preferential characteristics after training in
some domains. A system based on an associative memory may share these advantages
over a system based on backpropagation.
? Scalability: Many issues concerning the scalability of backpropagation networks
remain unresolved. It is not simple to build backpropagation networks of thousands
or hundreds of thousands of units. In contrast, Kanerva's Sparse Distributed Memory is specifically designed for such massive construction; one implementation on the
Connection Machine can contain l,O(XM)()() hidden units. The Genetic Memory shares
this property.
? Unity: Davis' algorithm has two levels of processing. The first level consists of
standard backpropagation networks, and the second is a meta-level which manipulates these networks. The Genetic Memory has incorporated both algorithms into a
single network; both algorithms are operating simultaneously.
My intuition is that different algorithms may be best suited for the different layers
of a neural network. Layers with a large fan-out (such as the input layer to the layer of hidden units) may be best driven by an algorithm suited to high-dimensional
searching, such as Genetic Algorithms or a Kohonen-style self-organizing system.
Layers with a large fan-in (such as the hidden-unit layer to the output layer) may be
best driven by a hill-climbing algorithms such a backpropagation.
CONCLUSIONS
? Real-world problems are often "high-dimensional", that is, are described by large
numbers of dependent variables. Algorithms must be specifically designed to function well in such high-dimensional spaces. Genetic Memory is such an algorithm .
? Genetic Memory, while sharing some features with Davis' approach, has fundamental differences that may make it more appropriate to some problems and easier to
scale to extremely-large (> 100,000 node) systems.
? The incorporation of the Genetic Algorithm improves the recall performance of a
standard associative memory.
? The structure of the Genetic Memory allows the user to access the parameters discovered by the Genetic Algorithm and used to assist in making the associations
stored in the memory.
463
464
Rogers
Acknowledgments
This work was supported in part by Cooperative Agreements NCC 2-408 and NCC
2-387 from the National Aeronautics and Space Administration (NASA) to the Universities Space Research Association (USRA). Funding related to the Connection
Machine was jointly provided by NASA and the Defense Advanced Research Projects
Agency (DARPA). All agencies involved were very helpful in promoting this
work, for which I am grateful.
The entire RIACS staff and the SDM group has been supportive of my work. Bruno
Olshausen was a vital sounding-board. Pentti Kanerva trusted my intuitions even
when the payoff wasn't yet clear. And finally, thanks to Doug Brockman, who decided to wait for me.
References
Davis, L., Genetic algoritluns and simulated annealing. London, England: Pitman
Publishing (1987).
Holland, J. H., Adaptation in natural and artificial systems. Ann Arbor: University of Michigan Press (1975).
Holland, J. H., "Escaping brittleness: the possibilities of general-purpose learning
algorithms applied to parallel rule-based systems," in Machine learning. an
artificial intelligence approach. Volume II, R. J. Michalski, J. G. Carbonell,
and T. M. Mitchell, eds. Los Altos, California: Morgan Kaufmann (1986).
Kanerva, Pentti., "Self-propagating Search: A Unified Theory of Memory," Center
for the Study of Language and Infonnation Report No. CSLI-84-7 (1984).
Kanerva, Pentti., Sparse distributed memory. Cambridge. Mass: MIT Press, 1988.
Rogers. David, "Using data-tagging to improve the perfonnance of Kanerva's sparse
distributed memory," Research Institute for Advanced Computer Science
Technical Report 88.1, NASA Ames Research Center (l988a).
Rogers, David, "Kanerva's Sparse Distributed Memory: an Associative Memory Algorithm Well-Suited to the Connection Machine," Int. J. High-Speed Comput., 2, pp. 349-365 (1989).
Rogers, David, "Statistical Prediction with Kanerva's Sparse Distributed Memory,"
Advances in Neural Information Processing Systems I, San Mateo: MorganKaufman (1989).
| 269 |@word middle:1 advantageous:1 open:2 pressure:4 initial:2 contains:2 score:1 selecting:1 genetic:60 interestingly:1 past:1 si:1 yet:1 written:2 must:2 riacs:1 designed:5 v:1 intelligence:1 selected:12 fewer:1 coarse:1 node:4 ames:2 location:44 preference:1 mathematical:2 along:1 consists:1 tagging:1 dist:1 ol:1 inspired:3 begin:1 discover:3 competes:1 underlying:1 provided:1 project:1 alto:1 mass:1 what:1 string:10 unified:1 every:3 ti:4 chronological:1 unit:10 local:2 switching:1 analyzing:1 black:3 mateo:1 dynamically:2 co:1 discriminatory:1 range:1 decided:2 acknowledgment:1 responsible:1 testing:1 backpropagation:9 procedure:5 jan:2 crossover:7 weather:27 word:1 refers:1 wait:1 selection:3 storage:4 writing:2 center:3 starting:1 assigns:1 manipulates:1 rule:2 continued:1 insight:1 brittleness:1 his:3 population:8 handle:1 searching:2 embedding:1 construction:1 user:3 massive:1 us:3 agreement:1 element:2 predicts:1 cooperative:1 cloud:4 quicker:1 worst:1 thousand:3 region:1 connected:1 cycle:2 counter:10 removed:2 incremented:1 intuition:2 environment:1 agency:2 grateful:1 segment:1 predictive:3 recombine:2 sep:1 darpa:1 represented:1 train:1 distinct:1 describe:1 effective:2 london:1 artificial:3 csli:1 larger:1 otherwise:1 ability:4 unseen:1 jointly:1 itself:1 associative:7 sdm:9 sequence:1 advantage:1 net:1 michalski:1 reconstruction:1 interaction:1 unresolved:1 adaptation:1 kohonen:1 beneath:2 organizing:1 poorly:2 scalability:2 los:1 parent:6 depending:2 propagating:1 augmenting:1 measured:1 aug:1 predicted:1 judge:1 australian:3 direction:3 radius:3 functionality:1 rogers:11 preliminary:1 recombined:2 predictiveness:2 purpose:1 wet:2 infonnation:1 sensitive:7 individually:1 create:4 reflects:1 trusted:1 mit:1 always:1 desirability:1 reconfigured:2 rather:1 june:1 improvement:1 greatly:1 contrast:2 am:1 helpful:1 dependent:1 entire:2 hidden:14 issue:1 denoted:1 summed:1 field:12 once:1 equal:4 represents:1 look:1 future:1 report:2 decremented:1 brockman:1 few:1 primarily:1 opening:1 preserve:1 simultaneously:1 national:1 uta:1 fitness:8 replacement:1 flxed:1 attempt:1 highly:6 possibility:1 weakness:1 held:2 implication:1 closer:1 preferential:1 perfonnance:1 unless:2 initialized:1 desired:1 increased:1 cover:4 measuring:1 applicability:1 addressing:1 entry:1 hundred:1 successful:2 conducted:1 stored:4 my:5 combined:1 ju:1 thanks:1 twelve:1 sensitivity:4 fundamental:1 together:1 reflect:2 central:1 possibly:1 book:1 style:2 toy:1 int:1 register:1 depends:1 piece:1 performed:3 view:2 wind:4 analyze:2 capability:1 parallel:1 jul:1 air:1 kaufmann:1 characteristic:1 who:1 ensemble:1 dry:3 climbing:1 raw:1 none:1 ncc:2 sharing:1 ed:1 against:1 pp:1 involved:1 hamming:3 sampled:1 newly:1 mitchell:1 recall:1 knowledge:1 improves:1 dimensionality:1 ou:1 nasa:4 feed:1 day:2 improved:1 formulation:1 done:4 box:3 mar:2 correlation:3 until:1 nonlinear:1 gray:1 olshausen:1 effect:2 hypothesized:1 contain:2 consisted:1 illustrated:1 during:3 self:2 davis:6 noted:1 m:1 hill:1 performs:1 l1:1 temperature:4 coast:2 discovers:2 funding:1 specialized:1 physical:3 insensitive:1 volume:1 million:1 association:2 cambridge:1 bruno:1 language:1 had:2 access:5 entail:1 operating:1 etc:2 add:1 align:1 feb:2 aeronautics:1 driven:2 termed:1 meta:1 binary:6 success:3 supportive:1 scoring:1 seen:2 minimum:1 greater:2 additional:1 staff:1 morgan:1 dew:1 rained:4 maximize:1 period:2 july:2 ii:2 multiple:2 desirable:4 technical:1 england:1 offer:2 concerning:1 coded:4 prediction:5 scalable:1 variant:1 ui1:1 dec:1 addressed:1 annealing:1 completes:1 envelope:1 operate:1 unlike:1 file:2 sounding:1 member:14 incorporates:1 call:1 vital:2 enough:1 iii:5 fit:2 architecture:2 associating:1 escaping:1 wasn:1 administration:1 whether:3 defense:1 assist:1 ul:1 forecasting:1 prefers:1 ignored:1 useful:1 clear:2 unimportant:2 detailed:1 ten:2 dna:3 percentage:1 write:1 group:1 four:5 threshold:1 changing:2 utilize:1 kept:1 ram:1 graph:3 groundbreaking:1 year:6 sum:3 powerful:1 place:2 reader:1 splicing:1 architectural:1 scaling:2 bit:17 layer:30 datum:2 fan:2 encountered:1 strength:1 occur:1 scanned:1 incorporation:1 speed:2 extremely:3 relatively:3 combination:3 remain:2 unity:1 evolves:1 making:2 temp:2 restricted:1 taken:2 kanerva:13 previously:3 turn:1 needed:1 end:2 studying:1 operation:3 promoting:1 appropriate:1 occurrence:1 robustness:1 original:1 rain:12 running:1 completed:2 publishing:1 opportunity:1 giving:1 recombination:4 build:1 february:3 upcoming:1 move:1 unclear:1 evolutionary:3 distance:6 separate:1 columnwise:1 simulated:1 me:2 carbonell:1 collected:1 length:1 code:3 fmding:1 difficult:2 anal:1 implementation:1 proper:1 perform:1 inspect:1 january:3 payoff:1 looking:1 incorporated:1 discovered:2 station:1 august:1 nwnber:1 david:4 connection:9 california:1 hour:9 nip:1 address:26 able:1 xm:1 reading:3 genetically:1 unsuccessful:1 memory:89 including:1 power:1 reconfigures:1 hybrid:8 difficulty:1 predicting:12 ranked:2 natural:1 advanced:3 older:1 improve:1 rated:6 l10:1 created:3 doug:1 jun:1 usra:1 prior:1 determining:1 relative:3 fully:1 expect:1 proportional:2 moffett:1 bulb:4 storing:1 share:2 row:3 course:1 changed:4 placed:2 surprisingly:1 last:1 supported:1 perceptron:1 institute:2 taking:1 absolute:1 sparse:17 pitman:1 distributed:17 dimension:1 world:5 evaluating:1 forward:1 collection:1 commonly:1 san:1 grater:1 far:1 correlate:2 reconstructed:1 nov:1 preferred:2 transcription:2 search:5 why:1 nature:1 ca:1 domain:5 apr:2 elation:1 noise:1 allowed:1 ule:1 fmd:1 site:1 referred:1 board:1 position:1 comput:1 candidate:1 reconfigure:4 bad:1 specific:2 swns:1 effectively:1 importance:1 demand:3 easier:1 suited:3 michigan:1 simply:1 likely:1 contained:2 holland:10 distrmuted:1 relies:1 oct:1 month:14 ann:1 towards:4 replace:1 content:1 change:4 specifically:3 determined:2 averaging:1 called:2 pentti:3 arbor:1 select:3 internal:1 latter:1 absolutely:1 tested:1 |
1,861 | 2,690 | On-Chip Compensation of Device-Mismatch
Effects in Analog VLSI Neural Networks
Miguel Figueroa
Department of Electrical Engineering, Universidad de Concepci?on
Casilla 160-C, Correo 3, Concepci?on, Chile
[email protected]
Seth Bridges and Chris Diorio
Computer Science & Engineering, University of Washington
Box 352350, Seattle, WA 98195-2350, USA
{seth, diorio}@cs.washington.edu
Abstract
Device mismatch in VLSI degrades the accuracy of analog arithmetic
circuits and lowers the learning performance of large-scale neural networks implemented in this technology. We show compact, low-power
on-chip calibration techniques that compensate for device mismatch. Our
techniques enable large-scale analog VLSI neural networks with learning performance on the order of 10 bits. We demonstrate our techniques
on a 64-synapse linear perceptron learning with the Least-Mean-Squares
(LMS) algorithm, and fabricated in a 0.35?m CMOS process.
1
Introduction
Modern embedded and portable electronic systems operate in unknown and mutating environments, and use adaptive filtering and machine learning techniques to discover the statistics of the data and continuously optimize their performance. Artificial neural networks
are an attractive substrate for implementing these techniques, because their regular computation and communication structures makes them a good match for custom VLSI implementations. Portable systems operate under severe power dissipation and space constraints,
and VLSI implementations provide a good tradeoff between computational throughput and
power/area cost. More specifically, analog VLSI neural networks perform their computation using the physical properties of transistors with orders of magnitude less power and
die area than their digital counterparts. Therefore, they could enable large-scale real-time
adaptive signal processing systems on a single die with minimal power dissipation.
Despite the promises delivered by analog VLSI, an important factor has prevented the success of large-scale neural networks using this technology: device mismatch. Gradients in
the parameters of the fabrication process create variations in the physical properties of silicon devices across a single chip. These variations translate into gain and offset mismatches
in the arithmetic blocks, which severely limit the overall performance of the system. As
a result, the accuracy of analog implementations rarely exceeds 5-6 bits, even for smallscale networks. This limitation renders these implementations useless for many important
applications. Although it is possible to combat some of these effects using careful design
techniques, they come at the cost of increased power and area, making an analog solution
less attractive.
(a) Single-layer LMS perceptron.
(b) Block diagram for the synapse.
Figure 1: A single-layer perceptron and synapse. (a) The output z of the perceptron is the inner
product between the input and weight vectors. The LMS algorithm updates the weights based on the
inputs and an error signal e. (b) The synapse stores the weight in an analog memory cell. A Gilbert
multiplier computes the product between the input and the weight and outputs a differential current.
The LMS block updates the weight.
We have built a 64-synapse analog neural network with an learning performance of 10 bits,
representing an improvement of more than one order of magnitude over that of traditional
analog designs, with a modest increase in power and die area. We fabricated our network
using a double-poly, 4-metal 0.35?m CMOS process available from MOSIS. We achieve
this performance by locally calibrating the critical analog blocks after circuit fabrication
using a combination of one-time (or periodic) and continuous calibration using the same
feedback as the network?s learning algorithm. We chose the Least Mean Squares (LMS) algorithm because of its simplicity and wide applicability in supervised learning techniques
such as adaptive filtering, adaptive inverse control, and noise canceling. Moreover, several useful unsupervised-learning techniques, such as adaptive orthogonalization, principal components analysis (PCA), independent components analysis (ICA) and decisionfeedback learning, use simple generalizations of LMS.
2
A linear LMS perceptron
Fig. 1(a) shows our system architecture, a linear perceptron with scalar output that performs
the function:
N
X
z(i) = bw0 (i) +
xj (i) wj (i)
(1)
j=1
where i represents time, z(i) is the output, xj (i) are the inputs, wj (i) are the synaptic
weights, and b is a constant bias input. We clarify the role of b in Section 3.1. After each
presentation of the input, the LMS algorithm updates the weights using the learning rule:
wj (i + 1) = wj (i) + ? xj (i) e(i)
i = 0 . . . N, x0 (i) = b
(2)
where ? is a constant learning rate, and e(i) is the error between the output and a reference
signal r(i) such that e(i) = r(i) ? z(i).
3
The synapse
Fig. 1(b) shows a block diagram of our synapse. We store the synaptic weights in a memory
cell that implements nonvolatile analog storage with linear updates. A circuit transforms
the single-ended voltage output of the memory cell (Vw ) into a differential voltage signal
(Vw+ , Vw? ), with a constant common mode. A Gilbert multiplier computes the 4-quadrant
product between this signal and the input (also represented as a differential voltage Vx+ ,
Vx? ). The output is a differential analog current pair (Io+ , Io? ), which we sum across all
synapses by connecting them to common wires.
(a) Measured output vs. input value.
(b) Measured output vs. weight value.
Figure 2: Gilbert multiplier response for 8 synapses.(a) Our multiplier maximizes the linearity of
xi , achieving a linear range of 600mV differential. Gain mismatch is 2:1 and offset mismatch is up to
200mV. (b) Our multiplier maximizes weight range at the cost of weight linearity (1V single-ended,
2V differential). The gain variation is lower, but the offset mismatch exceeds 60% of the range.
Because we represent the perceptron?s output and the reference with differential currents,
we can easily compute the error using simple current addition. We then transform (off-chip
in our current implementation) the resulting analog error signal using a pulse-density modulation (PDM) representation [1]. In this scheme, the value of the error is represented as
the difference between the density (frequency) of two fixed-width, fixed-amplitude digital
pulse trains (Pe+ and Pe? in Fig. 1(b)). These properties make the PDM representation
largely immune to amplitude and jitter noise. The performance of the perceptron is highly
sensitive to the resolution of the error signal; therefore the PDM representation is a good
match for it. The LMS block in the synapse takes the error and input values and computes
update pulses (also using PDM) according to Eqn. 2.
In the rest of this section, we analyze the effects of device mismatch in the performance
of the major blocks, discuss their impact in overall system performance, and present the
techniques that we developed to deal with them. We illustrate with experimental results
taken from silicon implementation of the perceptron in a 0.35?m CMOS process. All data
presented in this paper, unless otherwise stated, comes from this silicon implementation.
3.1 Multiplier
A Gilbert multiplier implements a nonlinear function of the product between two differential voltages. Device mismatch in the multiplier has two main effects: First, it creates
offsets in the inputs. Second, mismatch across the entire perceptron creates variations in
the offsets, gain, and linearity of the product. Thus, Eqn. 1 becomes:
N
X
z(i) =
aj fjx xj (i) ? dxj fjw wj (i) ? dw
x0 (i) = b
(3)
j
j=0
where aj represents the gain mismatch between multipliers, fjx and fjw are the nonlinearities applied to the inputs and weights (also mismatched across the perceptron), and dxj and
dw
j are the mismatched offsets of the inputs and weights.
Our analysis and simulations of the LMS algorithm [2] determine that the performance of
the algorithm is much more sensitive to the linearity of fjx than to the linearity of fjw , because the inputs vary over their dynamic range with a large bandwidth, while the bandwidth
of the weights is much lower than the adaptation time-constant. Therefore, the adaptation
compensates for mild nonlinearities in the weights as long as fjw remains a monotonic odd
function [2]. Consequently, we sized the transistors in the Gilbert multiplier to maximize
the linearity of fjx , but paid less attention (in order to minimize size and power) to fjw .
Fig. 2(a) shows the output of 8 synapses in the system as a function of the input value. The
(a) Memory cell circuit.
(b) Measured weight updates
Figure 3: A simple PDM analog memory cell. (a) We store each weight as nonvolatile analog charge
on the floating gate FG. The weight increments and decrements are proportional to the density of the
pulses on Pinc and Pdec . (b) Memory updates as a function of the increment and decrement pulse
densities for 8 synapses. The updates show excellent linearity (10 bits), but also poor matching both
within a synapse and between synapses.
response is highly linear. The gain mismatch is about 2:1, but the LMS algorithm naturally
absorbs it into the learned weight value. Fig. 2(b) shows the multiplier output as a function
of the single-ended weight value Vw . The linearity is visibly worse in this case, but the
LMS algorithm compensates for it.
The graphs in the Fig. 2 also show the input and weight offsets. Because of the added
mismatch in the single-ended to differential converter, the weights present an offset of up
to ?300mV, or 30% of the weight range. The LMS algorithm will also compensate for this
offset by absorbing it into the weight, as shown in the analysis of [3] for backprogagation
neural networks. However, this will only occur if the weight range is large enough accomodate for the offset mismatch. Consequently, we sacrifice weight linearity to increase
the weight range. Input offsets pose a harder problem, though. The offsets are small (up
100mV), but because of the restricted input range (to maximize linearity), they are large
enough to dramatically affect the learning performance of the perceptron. Our solution was
to use the bias synapse w0 to compensate for the accumulated input offset. Assuming that
the multiplier is linear, offsets translate into nonzero-mean inputs, which a bias synapse
trained with LMS can remove as demonstrated in [4]. To guarantee sufficient gain, we
provide a stronger bias current to the multiplier in the bias synapse.
3.2 Memory cell
A synapse transistor [5] is a silicon device that provides compact, accurate, nonvolatile
analog storage as charge on its floating gate. Fowler-Nordheim tunneling adds charge to
the floating gate and hot-electron injection removes charge. Both mechanisms can be used
to accurately update the stored value during normal device operation. Because of these
properties, synapse transistors have been a popular choice for weight storage in recent
silicon learning systems [6, 7].
Despite the advantages listed above, it is hard to implement linear learning rules such as
LMS using tunneling and injection. This is because their dynamics are exponential with respect to their control variables (floating-gate voltage, tunneling voltage and injection drain
current), which naturally lead to weight-dependent nonlinear update rules. This is an important problem because the learning performance of the perceptron is strongly dependent
on the accuracy of the weight updates; therefore distortions in the learning rule will degrade
performance. The initial design of our memory cell, shown in Fig. 3(a) and based on the
work presented in [8], solves this problem: We store the analog weight as charge on the
floating gate FG of synapse transistor M1 . Pulses on Pdec and Pinc activate tunneling and
injection and add or remove charge from the floating gate, respectively. The operational
(a) Calibrated memory cell circuit.
(b) Measured calibrated weight updates.
Figure 4: PDM memory cell with local calibration. (a) We first match the tunneling rate across all
synapses by locally changing the voltage at the floating gate FGdec . Then, we modify the injection
rate to match the local tunneling rate using the floating gate FGinc . (b) The calibrated updates are
symmetric and uniform within 9-10 bits.
amplifier sets the floating-gate voltage at the global voltage Vbias . Capacitor Cw integrates
the charge updates, changing the output Vout by ?Vout = ?Q/C. Because the floatinggate voltage is constant and so are the pulse widths and amplitudes, the magnitude of the
updates depends on the density of the pulses Pinc and Pdec . Fig. 3(b) shows the magnitude
of the weight updates as a function of the density of pulses in Pinc (positive slopes) and
Pdec (negative slopes) for 8 synapses. The linearity of the updates, measured as the integral
nonlinearity (INL) of the transfer functions depicted in Fig. 3(b), exceeds 10 bits.
Fig. 3(b) highlights an important problem caused by device mismatch: the strengths of
tunneling and injection are poorly balanced within a synapse (the slopes show up to a 4:1
mismatch). Moreover, they show a variation of more than 3:1 across different synapses
in the perceptron. This translates into asymmetric update rules that are also nonuniform
across synapses. The local asymmetry of the learning rate translates into offsets between
the learned and target weights, degrading the learning performance of the perceptron. The
nonuniformity between learning rates across the perceptron changes Eqn. 2 into:
wj (i + 1) = wj (i) + ?j xj (i) e(i)
i = 0 . . . N, x0 (i) = b
(4)
where ?j are the different learning rates for each synapse. Generalizing the conventional
stability analysis of LMS [9], we can show that the condition for the stability of the weight
vector is: 0 < ?max < 1/?max , where ?max is the maximal eigenvalue of the input?s
correlation matrix and ?max = maxj (?j ). Therefore, learning rate mismatch does not
affect the accuracy of the learned weights, but it does slow down convergence because we
need to scale all learning rates globally to limit the value of the maximal rate.
To maintain good learning performance and convergence speed, we need to make learning
rates symmetric and uniform across the perceptron. We modified the design of the memory
cell to incorporate local calibration mechanisms that achieve this goal. Fig. 4(a) shows our
new design. The first step is to equalize tunneling rates: The voltage at the new floating gate
FGdec sets the voltage at the floating-gate FG and controls the ratio between the strength
of tunneling and injection onto FG: Raising the voltage at FGdec increases the drain-tochannel voltage and reduces the gate-to-tunneling-junction voltage at M1 , thus increasing
injection efficiency and reducing tunneling strength [5]. We set the voltage at FGdec by first
tunneling using the global line erase dec, and then injecting on transistor M3 by lowering
the local line set dec to equalize the tunneling rates across all synapses. To compare the
tunneling rates, we issue a fixed number of pulses at Pdec and compare the memory cell
outputs using a double-sampling comparator (off-chip in the current implementation). To
control the injection rate, we add transistor M2 , which limits the current through M1 and
(a) LMS block.
(b) Measured RMS error.
Figure 5: LMS block at each synapse. (a) The difference between the densities of Pinc and Pdec
is proportional to the product between the input and the error, and thus constitutes an LMS update
rule. (b) RMS error for a single-synapse with a constant input and reference, including a calibrated
memory cell with symmetric updates, a simple synapse with asymmetric updates, and a simulated
ideal synapse.
thus the injection strength of the pulse at Pinc . We control the current limit with the voltage
at the new floating gate FGinc : we first remove electrons from the floating gate using the
global line erase inc. Then we inject on transistor M4 by lowering the local line set inc to
match the injection rates across all synapses. The entire process is controlled by a simple
state machine (also currently off-chip). Fig. 4(b) shows the tunneling and injection rates
after calibration as a function of the density of pulses Pinc and Pdec . Comparing the graph
to Fig. 4(b), it is clear that the update rates are now symmetric and uniform across all
synapses (they match within 9-10 bits). Note that we could also choose to calibrate just for
learning rate symmetry and not uniformity across synapses, thus eliminating the floating
gate FGinc and its associated circuitry. This optimization would result in approximately a
25% reduction in memory cell area (6% reduction in total synapse area), but would also
cause an increase of more than 200% in convergence time, as illustrated in Section 4.
3.3 The LMS block
Fig. 5(a) shows a block diagram of the LMS-update circuit at each synapse. A pulsedensity modulator [10] transforms the synaptic input into a pair of digital pulse-trains of
?
fixed width (P+
x , Px ). The value of the input is represented as the difference between the
density (frequency) of the pulse trains. We implement the memory updates of Eqn. 2 by
?
digitally combining the input and error pulses (P+
e , Pe ) such that:
Pinc
Pdec
= (Px+ AN D Pe+ ) OR (Px? AN D Pe? )
= (Px+ AN D Pe? ) OR (Px? AN D Pe+ )
(5)
(6)
This technique was used previously in a synapse-transistor based circuit that learns correlations between signals [11], and to multiply and add signals [1]. If the pulse trains are
asynchronous and sparse, then using Eqn. 5 and Eqn. 6 to increment and decrement the
synaptic weight implements the LMS learning rule of Eqn. 2.
To validate our design, we first trained a single synapse with a DC input to learn a constant
reference. Because the input is constant, the linearity and offsets in the input signal do not
affect the learning performance; therefore this experiment tests the resolution of the feedback path (LMS circuit and memory cell) isolated from the analog multipliers. Fig. 5(b)
shows the evolution of the RMS value of the error for a synapse using the original and
calibrated memory cells. The resolution of the pulse-density modulators is about 8 bits,
which limits the resolution of the error signal. We also show the RMS error for a simulated (ideal) synapse learning from the same error. We plot the results in a logarithmic
scale to highlight the differences between the three curves. The RMS error of the calibrated synapse converges to about 0.1nA. Computing the equivalent resolution in bits as
(a) Measured RMS error.
(b) Measured weight evolution.
Figure 6: Results for 64-synapse experiment. (a) Asymmetric learning rates and multiplier offsets
limit the output resolution to around 3 bits. Symmetric learning rates and a bias synapse brings
the resolution up to more 10 bits, and uniform updates reduce convergence time. (b) Synapse 4
shows a larger mismatch than synapse 1 and therefore it deviates from its theoretical target value
to compensate. The bias synapse in the VLSI perceptron converges to a value that compensates for
offsets in the inputs xi to the multipliers.
RM S error
rb = ?log2 0.5 output
, we find that for a 2?A output range, this error represents an
range
output resolution of about 13 bits. The difference with the simulated synapse is due to the
discrete weight updates in the PDM memory cell. Without calibration, the RMS error converges to 0.4nA (or about 11 bits), due to the offset in the learned weights introduced by the
asymmetry in the learning rate. As discussed in Section 4, the degradation of the learning
performance in a larger-scale system due to asymmetric learning rates is drastically larger.
4
A 64-synapse perceptron
To test our techniques in a larger-scale system, we fabricated a 64-synapse linear perceptron in a 0.35?m CMOS process. The circuit uses 0.25mm2 of die area and dissipates
200?W. Fig. 6(a) shows the RMS error of the output in a logarithmic scale as we introduce
different compensation techniques. We used random zero-mean inputs selected from a uniform distribution over the entire input range, and trained the network using the response
from a simulated perceptron with ideal multipliers and fixed weights as a reference. In
our first experiments, we trained the network without using any compensation. The error
settles to 10?A RMS, which corresponds to an output resolution of about 3 bits for a full
range of 128?A differential. Calibrating the synapses for symmetric learning rates only
improves the RMS error to 5?A (4 bits), but the error introduced by the multiplier offsets
still dominates the residual error. Introducing the bias synapse and keeping the learning
rates symmetric (but nonuniform across the perceptron) compensates for the offsets and
brings the error down to 60nA RMS, corresponding to an output resolution better than 10
bits. Further calibrating the synapses to achieve uniform, symmetric learning rates maintains the same learning performance, but reduces convergence time to less than one half, as
predicted by the analysis in Section 3.2. A simulated software perceptron with ideal multipliers and LMS updates that uses an error signal of the same resolution as our experiments
gives an upper bound of just under 12 bits for the learning performance.
Fig. 6(b) depicts the evolution of selected weights in the silicon perceptron with on-chip
compensation and the software version. The graph shows that synapse 1 in our VLSI implementation suffers from little mismatch, and therefore its weight virtually converges to
the theoretical value given by the software implementation. Because the PDM updates are
discrete, the weight shows a larger oscillation around its target value than the software version. Synapse 4 shows a larger mismatch; therefore it converges to a visibly different value
from the theoretical in order to compensate for it. The bias weight in the software percep-
tron converges to zero because the inputs have zero mean. In the VLSI perceptron, input
offsets in the multipliers create nonzero-mean inputs; therefore the bias synapse converges
to a value that compensates for the aggregated effect of the offsets. The normalized value
of -1.2 reflects the gain boost given to this multiplier to increase its dynamic range.
5
Conclusions
Device mismatch prevents analog VLSI neural networks from delivering good learning
performance for large-scale applications. We identified the key effects of mismatch and
presented on-chip compensation techniques. Our techniques rely both on one-time (or
periodic) calibration, and on the adaptive operation of the system to achieve continuous
calibration. Combining these techniques with careful circuit design enables an improvement of more than one order of magnitude in accuracy compared to traditional analog
designs, at the cost of an off-line calibration phase and a modest increase in die area and
power. We illustrated our techniques with a 64-synapse analog-VLSI linear perceptron that
adapts using the LMS algorithm. Future work includes extending these techniques to unsupervised learning algorithms such as adaptive orthogonalization, principal components
analysis (PCA) and independent components analysis (ICA).
Acknowledgements
This work was financed in part by the Chilean government through FONDECYT grant
#1040617. We fabricated our chips through MOSIS.
References
[1] Y. Hirai and K. Nishizawa, ?Hardware implementation of a PCA learning network by an asynchronous PDM digital circuit,? in IEEE-INNS-ENNS International Joint Conference on Neural
Networks (IJCNN), vol. 2, pp. 65?70, 2000.
[2] M. Figueroa, Adaptive Signal Processing and Correlational Learning in Mixed-Signal VLSI.
Ph.D. Thesis, University of Washington, 2005.
[3] B. K. Dolenko and H. C. Card, ?Tolerance to analog hardware of on-chip learning in backpropagation networks,? IEEE Transactions on Neural Networks, vol. 6, no. 5, pp. 1045?1052,
1995.
[4] F. Palmieri, J. Zhu, and C. Chang, ?Anti-Hebbian learning in topologically constrained linear
networks: A tutorial,? IEEE Transactions on Neural Networks, vol. 4, no. 5, pp. 748?761, 1993.
[5] C. Diorio, P. Hasler, B. Minch, and C. Mead, ?A complementary pair of four-terminal silicon
synapses,? Analog Integrated Circuits and Signal Processing, vol. 13, no. 1/2, pp. 153?166,
1997.
[6] C. Diorio, D. Hsu, and M. Figueroa, ?Adaptive CMOS: from biological inspiration to systemson-a-chip,? Proceedings of the IEEE, vol. 90, no. 3, pp. 345?357, 2002.
[7] J. Dugger and P. Hasler, ?Improved correlation learning rule in continuously adapting floatinggate arrays using logarithmic pre-distortion of input and learning signals,? in IEEE Intl. Symposium on Circuits and Systems (ISCAS), vol. 2, pp. 536?539, 2002.
[8] C. Diorio, S. Mahajan, P. Hasler, B. A. Minch, and C. Mead, ?A high-resolution nonvolatile
analog memory cell,? in IEEE Intl. Symp. on Circuits and Systems, vol. 3, pp. 2233?2236,
1995.
[9] B. Widrow and E. Walach, Adaptive Inverse Control. Upper Saddle River, NJ: Prentice-Hall,
1996.
[10] C. Mead, Analog VLSI and Neural Systems. Reading, MA: Addison-Wesley, 1989.
[11] A. Shon, D. Hsu, and C. Diorio, ?Learning spike-based correlations and conditional probabilities in silicon,? in Neural Information Processing Systems (NIPS), (Vancouver, BC), 2001.
| 2690 |@word mild:1 version:2 eliminating:1 stronger:1 pulse:17 simulation:1 paid:1 harder:1 reduction:2 initial:1 bc:1 current:10 comparing:1 percep:1 enables:1 remove:4 plot:1 update:28 v:2 half:1 selected:2 device:11 floatinggate:2 chile:1 provides:1 differential:10 symposium:1 absorbs:1 symp:1 introduce:1 x0:3 sacrifice:1 ica:2 terminal:1 globally:1 little:1 increasing:1 becomes:1 erase:2 discover:1 linearity:12 moreover:2 circuit:14 maximizes:2 degrading:1 developed:1 fabricated:4 ended:4 nj:1 guarantee:1 combat:1 charge:7 smallscale:1 rm:1 control:6 fjw:5 grant:1 positive:1 engineering:2 local:6 modify:1 limit:6 severely:1 io:2 despite:2 mead:3 path:1 modulation:1 approximately:1 chose:1 range:13 block:11 implement:5 backpropagation:1 area:8 adapting:1 matching:1 pre:1 regular:1 quadrant:1 onto:1 storage:3 prentice:1 optimize:1 gilbert:5 conventional:1 demonstrated:1 equivalent:1 attention:1 resolution:12 simplicity:1 m2:1 rule:8 array:1 dw:2 stability:2 variation:5 increment:3 target:3 substrate:1 us:2 asymmetric:4 role:1 electrical:1 vbias:1 wj:7 diorio:6 balanced:1 equalize:2 environment:1 digitally:1 pdm:9 dynamic:3 trained:4 uniformity:1 creates:2 efficiency:1 easily:1 seth:2 joint:1 chip:11 represented:3 train:4 activate:1 modulators:1 artificial:1 larger:6 distortion:2 otherwise:1 compensates:5 statistic:1 transform:1 delivered:1 advantage:1 eigenvalue:1 transistor:9 inn:1 product:6 maximal:2 adaptation:2 combining:2 financed:1 translate:2 poorly:1 achieve:4 adapts:1 validate:1 seattle:1 convergence:5 double:2 asymmetry:2 extending:1 intl:2 cmos:5 converges:7 illustrate:1 widrow:1 pose:1 miguel:1 measured:8 odd:1 solves:1 implemented:1 c:1 predicted:1 come:2 vx:2 enable:2 settle:1 implementing:1 government:1 generalization:1 biological:1 clarify:1 around:2 hall:1 normal:1 lm:25 electron:2 circuitry:1 major:1 vary:1 integrates:1 injecting:1 currently:1 bridge:1 sensitive:2 create:2 reflects:1 modified:1 voltage:17 improvement:2 visibly:2 dependent:2 accumulated:1 entire:3 integrated:1 vlsi:14 enns:1 overall:2 issue:1 constrained:1 washington:3 sampling:1 mm2:1 represents:3 unsupervised:2 throughput:1 constitutes:1 future:1 modern:1 m4:1 floating:14 maxj:1 phase:1 iscas:1 maintain:1 amplifier:1 highly:2 multiply:1 custom:1 severe:1 accurate:1 integral:1 modest:2 unless:1 isolated:1 theoretical:3 minimal:1 increased:1 calibrate:1 cost:4 applicability:1 introducing:1 uniform:6 fabrication:2 stored:1 periodic:2 minch:2 calibrated:6 concepci:2 density:10 international:1 river:1 universidad:1 off:4 connecting:1 continuously:2 na:3 thesis:1 choose:1 worse:1 inject:1 nonlinearities:2 de:1 includes:1 inc:2 caused:1 mv:4 depends:1 analyze:1 maintains:1 slope:3 minimize:1 square:2 accuracy:5 largely:1 vout:2 accurately:1 synapsis:16 suffers:1 canceling:1 synaptic:4 frequency:2 pp:7 naturally:2 associated:1 gain:8 hsu:2 popular:1 palmieri:1 improves:1 amplitude:3 wesley:1 supervised:1 hirai:1 response:3 improved:1 synapse:41 box:1 though:1 strongly:1 just:2 correlation:4 eqn:7 nonlinear:2 mode:1 brings:2 aj:2 fowler:1 usa:1 effect:6 calibrating:3 normalized:1 multiplier:21 counterpart:1 evolution:3 inspiration:1 symmetric:8 nonzero:2 mahajan:1 illustrated:2 deal:1 attractive:2 during:1 width:3 die:6 demonstrate:1 tron:1 performs:1 dissipation:2 orthogonalization:2 nonuniformity:1 common:2 absorbing:1 physical:2 analog:26 discussed:1 m1:3 silicon:8 nonlinearity:1 immune:1 calibration:9 add:4 recent:1 store:4 success:1 determine:1 maximize:2 aggregated:1 signal:16 arithmetic:2 full:1 reduces:2 hebbian:1 exceeds:3 match:6 compensate:5 long:1 prevented:1 controlled:1 impact:1 represent:1 cell:17 dec:2 addition:1 diagram:3 operate:2 rest:1 virtually:1 dxj:2 capacitor:1 vw:4 ideal:4 enough:2 xj:5 affect:3 architecture:1 bandwidth:2 converter:1 modulator:1 inner:1 reduce:1 identified:1 tradeoff:1 translates:2 pca:3 rms:11 render:1 cause:1 dramatically:1 useful:1 clear:1 listed:1 delivering:1 transforms:2 locally:2 ph:1 hardware:2 tutorial:1 rb:1 discrete:2 promise:1 vol:7 key:1 four:1 achieving:1 changing:2 hasler:3 lowering:2 mosis:2 graph:3 sum:1 inverse:2 jitter:1 topologically:1 electronic:1 oscillation:1 tunneling:15 bit:17 layer:2 bound:1 strength:4 occur:1 figueroa:3 constraint:1 ijcnn:1 software:5 speed:1 injection:12 px:5 department:1 according:1 combination:1 poor:1 across:14 making:1 restricted:1 taken:1 remains:1 previously:1 discus:1 mechanism:2 addison:1 mutating:1 available:1 operation:2 junction:1 gate:15 original:1 log2:1 added:1 spike:1 degrades:1 traditional:2 gradient:1 cw:1 card:1 simulated:5 w0:1 chris:1 degrade:1 portable:2 assuming:1 useless:1 ratio:1 stated:1 negative:1 implementation:11 design:8 unknown:1 perform:1 upper:2 wire:1 compensation:5 anti:1 communication:1 dc:1 nonuniform:2 introduced:2 pair:3 raising:1 learned:4 boost:1 nordheim:1 nip:1 mismatch:23 reading:1 built:1 max:4 memory:19 including:1 power:9 critical:1 hot:1 rely:1 dissipates:1 residual:1 zhu:1 representing:1 scheme:1 technology:2 deviate:1 acknowledgement:1 drain:2 vancouver:1 embedded:1 highlight:2 mixed:1 limitation:1 filtering:2 proportional:2 digital:4 metal:1 sufficient:1 asynchronous:2 keeping:1 drastically:1 bias:10 perceptron:26 mismatched:2 wide:1 sparse:1 fg:4 tolerance:1 feedback:2 curve:1 computes:3 adaptive:10 transaction:2 compact:2 global:3 xi:2 continuous:2 learn:1 transfer:1 operational:1 symmetry:1 excellent:1 cl:1 poly:1 main:1 decrement:3 noise:2 complementary:1 fig:17 depicts:1 nonvolatile:4 slow:1 exponential:1 pe:7 learns:1 down:2 offset:23 dominates:1 magnitude:5 accomodate:1 depicted:1 generalizing:1 logarithmic:3 saddle:1 prevents:1 inl:1 scalar:1 shon:1 chang:1 monotonic:1 corresponds:1 ma:1 comparator:1 conditional:1 sized:1 presentation:1 goal:1 consequently:2 careful:2 hard:1 change:1 specifically:1 reducing:1 principal:2 degradation:1 total:1 correlational:1 experimental:1 m3:1 rarely:1 incorporate:1 |
1,862 | 2,691 | Hierarchical Distributed Representations for
Statistical Language Modeling
John Blitzer, Kilian Q. Weinberger, Lawrence K. Saul, and Fernando C. N. Pereira
Department of Computer and Information Science, University of Pennsylvania
Levine Hall, 3330 Walnut Street, Philadelphia, PA 19104
{blitzer,kilianw,lsaul,pereira}@cis.upenn.edu
Abstract
Statistical language models estimate the probability of a word occurring
in a given context. The most common language models rely on a discrete
enumeration of predictive contexts (e.g., n-grams) and consequently fail
to capture and exploit statistical regularities across these contexts. In this
paper, we show how to learn hierarchical, distributed representations of
word contexts that maximize the predictive value of a statistical language
model. The representations are initialized by unsupervised algorithms for
linear and nonlinear dimensionality reduction [14], then fed as input into
a hierarchical mixture of experts, where each expert is a multinomial distribution over predicted words [12]. While the distributed representations
in our model are inspired by the neural probabilistic language model of
Bengio et al. [2, 3], our particular architecture enables us to work with
significantly larger vocabularies and training corpora. For example, on a
large-scale bigram modeling task involving a sixty thousand word vocabulary and a training corpus of three million sentences, we demonstrate
consistent improvement over class-based bigram models [10, 13]. We
also discuss extensions of our approach to longer multiword contexts.
1
Introduction
Statistical language models are essential components of natural language systems for
human-computer interaction. They play a central role in automatic speech recognition [11],
machine translation [5], statistical parsing [8], and information retrieval [15]. These models estimate the probability that a word will occur in a given context, where in general
a context specifies a relationship to one or more words that have already been observed.
The simplest, most studied case is that of n-gram language modeling, where each word is
predicted from the preceding n?1 words. The main problem in building these models is
that the vast majority of word combinations occur very infrequently, making it difficult to
estimate accurate probabilities of words in most contexts.
Researchers in statistical language modeling have developed a variety of smoothing techniques to alleviate this problem of data sparseness. Most smoothing methods are based on
simple back-off formulas or interpolation schemes that discount the probability of observed
events and assign the ?leftover? probability mass to events unseen in training [7]. Unfortunately, these methods do not typically represent or take advantage of statistical regularities
among contexts. One expects the probabilities of rare or unseen events in one context to be
related to their probabilities in statistically similar contexts. Thus, it should be possible to
estimate more accurate probabilities by exploiting these regularities.
Several approaches have been suggested for sharing statistical information across contexts.
The aggregate Markov model (AMM) of Saul and Pereira [13] (also discussed by Hofmann
and Puzicha [10] as a special case of the aspect model) factors the conditional probability
table of a word given its context by a latent variable representing context ?classes?. However, this latent variable approach is difficult to generalize to multiword contexts, as the
size of the conditional probability table for class given context grows exponentially with
the context length.
The neural probabilistic language model (NPLM) of Bengio et al. [2, 3] achieved significant improvements over state-of-the-art smoothed n-gram models [6]. The NPLM encodes
contexts as low-dimensional continuous vectors. These are fed to a multilayer neural network that outputs a probability distribution over words. The low-dimensional vectors and
the parameters of the network are trained simultaneously to minimize the perplexity of the
language model. This model has no difficulty encoding multiword contexts, but its training
and application are very costly because of the need to compute a separate normalization for
the conditional probabilities associated to each context.
In this paper, we introduce and evaluate a statistical language model that combines the
advantages of the AMM and NPLM. Like the NPLM, it can be used for multiword contexts, and like the AMM it avoids per-context normalization. In our model, contexts are
represented as low-dimensional real vectors initialized by unsupervised algorithms for dimensionality reduction [14]. The probabilities of words given contexts are represented by
a hierarchical mixture of experts (HME) [12], where each expert is a multinomial distribution over predicted words. This tree-structured mixture model allows a rich dependency
on context without expensive per-context normalization. Proper initialization of the distributed representations is crucial; in particular, we find that initializations from the results
of linear and nonlinear dimensionality reduction algorithms lead to better models (with
significantly lower test perplexities) than random initialization.
In practice our model is several orders of magnitude faster to train and apply than the
NPLM, enabling us to work with larger vocabularies and training corpora. We present results on a large-scale bigram modeling task, showing that our model also leads to significant
improvements over comparable AMMs.
2
Distributed representations of words
Natural language has complex, multidimensional semantics. As a trivial example, consider
the following four sentences:
The vase broke.
The window broke.
The vase contains water.
The window contains water.
The bottom right sentence is syntactically valid but semantically meaningless. As shown by
the table, a two-bit distributed representation of the words ?vase? and ?window? suffices to
express that a vase is both a container and breakable, while a window is breakable but cannot be a container. More generally, we expect low dimensional continuous representations
of words to be even more effective at capturing semantic regularities.
Distributed representations of words can be derived in several ways. In a given corpus of
text, for example, consider the matrix of bigram counts whose element C
Pij records the
number of times that word wj follows word wi . Further, let pij = Cij / k Cik denote
the conditional frequencies derived from these counts, and let p~i denote the V -dimensional
frequency vector with elements pij , where V is the vocabulary size. Note that the vectors p~i
themselves provide a distributed representation of the words wi in the corpus. For large
vocabularies and training corpora, however, this is an extremely unwieldy representation,
tantamount to storing the full matrix of bigram counts. Thus, it is natural to seek a lower
dimensional representation that captures the same information. To this end, we need to map
each vector p~i to some d-dimensional vector ~xi , with d V . We consider two methods in
dimensionality reduction for this problem. The results from these methods are then used to
initialize the HME architecture in the next section.
2.1
Linear dimensionality reduction
The simplest form of dimensionality reduction is principal component analysis (PCA).
PCA computes a linear projection of the frequency vectors p~i into the low dimensional
subspace that maximizes their variance. The variance-maximizing subspace of dimensionality d is spanned by the top d eigenvectors of the frequency vector covariance matrix. The
eigenvalues of the covariance matrix measure the variance captured by each axis of the
subspace. The effect of PCA can also be understood as a translation and rotation of the
frequency vectors p~i , followed by a truncation that preserves only their first d elements.
2.2
Nonlinear dimensionality reduction
Intuitively, we would like to map the vectors p~i into a low dimensional space where semantically similar words remain close together and semantically dissimilar words are far
apart. Can we find a nonlinear mapping that does this better than PCA? Weinberger et al.
recently proposed a new solution to this problem based on semidefinite programming [14].
Let ~xi denote the image of p~i under this mapping. The mapping is discovered by first
2
learning the V ?V matrix of Euclidean squared distances [1] given by Dij = |~xi ? ~xj | .
This is done by balancing two competing goals: (i) to co-locate semantically similar words,
and (ii) to separate semantically dissimilar words. The first goal is achieved by fixing the
distances between words with similar frequency vectors to their original values. In particular, if p~j and p~k lie within some small neighborhood of each other, then the corresponding
2
element Djk in the distance matrix is fixed to the value |~
pj ? p~k | . The second goal is
achieved by maximizing the sum of pairwise squared distances ?ij Dij . Thus, we push the
words in the vocabulary as far apart as possible subject to the constraint that the distances
between semantically similar words do not change.
The only freedom in this optimization is the criterion for judging that two words are semantically similar. In practice, we adopt a simple criterion such as k-nearest neighbors in
the space of frequency vectors p~i and choose k as small as possible so that the resulting
neighborhood graph is connected [14].
The optimization is performed over the space of Euclidean squared distance matrices [1].
Necessary and sufficient conditions for the matrix D to be interpretable as a Euclidean
squared distance matrix are that D is symmetric and that the Gram matrix1 derived from
G = ? 12 HDH T is semipositive definite, where H = I ? V1 11T . The optimization can
thus be formulated as the semidefinite programming problem:
Maximize ?ij Dij subject to: (i) DT = D, (ii) ? 12 HDH 0, and
2
(iii) Dij = |~
pi ? p~j | for all neighboring vectors p~i and p~j .
1
Assuming without loss of generality that the vectors ~
xi are centered on the origin, the dot products Gij = ~
xi ? ~
xj are related to the pairwise squared distances Dij = |~
xi ? ~
xj |2 as stated above.
PCA
SDE
0.0
0.2
0.4
0.6
0.8
1.0
Figure 1: Eigenvalues from principal component analysis (PCA) and semide?nite embedding (SDE), applied to bigram distributions of the 2000 most frequently occuring words in
the corpus. The eigenvalues, shown normalized by their sum, measure the relative variance
captured by individual dimensions.
The optimization is convex, and its global maximum can be computed in polynomial
time [4]. The optimization here differs slightly from the one used by Weinberger et al. [14]
in that here we only preserve local distances, as opposed to local distances and angles.
After computing the matrix Dij by semide?nite programming, a low dimensional embedding ~xi is obtained by metric multidimensional scaling [1, 9, 14]. The top eigenvalues of
the Gram matrix measure the variance captured by the leading dimensions of this embedding. Thus, one can compare the eigenvalue spectra from this method and PCA to ascertain
if the variance of the nonlinear embedding is concentrated in fewer dimensions. We refer
to this method of nonlinear dimensionality reduction as semide?nite embedding (SDE).
Fig. 1 compares the eigenvalue spectra of PCA and SDE applied to the 2000 most frequent
words2 in the corpus described in section 4. The ?gure shows that the nonlinear embedding
by SDE concentrates its variance in many fewer dimensions than the linear embedding by
PCA. Indeed, Fig. 2 shows that even the ?rst two dimensions of the nonlinear embedding
preserve the neighboring relationships of many words that are semantically similar. By
contrast, the analogous plot generated by PCA (not shown) reveals no such structure.
MONDAY
TUESDAY
WEDNESDAY
THURSDAY
FRIDAY
SATURDAY
SUNDAY
MAY, WOULD, COULD, SHOULD,
MIGHT, MUST, CAN, CANNOT,
COULDN'T, WON'T, WILL
ONE, TWO, THREE,
FOUR, FIVE, SIX,
SEVEN, EIGHT, NINE,
TEN, ELEVEN,
TWELVE, THIRTEEN,
FOURTEEN, FIFTEEN,
SIXTEEN,
SEVENTEEN,
EIGHTEEN
MILLION
BILLION
ZERO
JANUARY
FEBRUARY
MARCH
APRIL
JUNE
JULY
AUGUST
SEPTEMBER
OCTOBER
NOVEMBER
DECEMBER
Figure 2: Projection of the normalized bigram counts of the 2000 most frequent words
onto the ?rst two dimensions of the nonlinear embedding obtained by semide?nite programming. Note that semantically meaningful neighborhoods are preserved, despite the
massive dimensionality reduction from V = 60000 to d = 2.
2
Though convex, the optimization over distance matrices for SDE is prohibitively expensive for
large matrices. For the results in this paper?on the corpus described in section 4?we solved the
semide?nite program in this section to embed the 2000 most frequent words in the corpus, then used
a greedy incremental solver to embed the remaining 58000 words in the vocabulary. Details of this
incremental solver will be given elsewhere. Though not the main point of this paper, the nonlinear
embedding of V = 60000 words is to our knowledge one of the largest applications of recently
developed spectral methods for nonlinear dimensionality reduction [9, 14].
3
Hierarchical mixture of experts
The model we use to compute the probability that word w0 follows word w is known as a
hierarchical mixture of experts (HME) [12]. HMEs are fully probabilistic models, making
them ideally suited to the task of statistical language modeling. Furthermore, like multilayer neural networks they can parameterize complex, nonlinear functions of their input.
Figure 3 depicts a simple, two-layer HME. HMEs are tree-structured mixture models in
which the mixture components are ?experts? that lie at the leaves of the tree. The interior
nodes of the tree perform binary logistic regressions on the input vector to the HME, and
the mixing weight for a leaf is computed by multiplying the probabilities of each branch
(left or right) along the path to that leaf. In our model, the input vector ~x is a function of
the context word w, and the expert at each leaf specifies a multinomial distribution over
the predicted word w0 . Letting ? denote a path through the tree from root to leaf, the HME
computes the probability of a word w0 conditioned on a context word w as
X
Pr(w0 |w) =
Pr(?|~x(w)) ? Pr(w0 |?).
(1)
?
We can compute the maximum likelihood parameters for the HME using an ExpectationMaximization (EM) algorithm [12]. The E-step involves computing the posterior probability over paths Pr(?|w, w0 ) for each observed bigram in the training corpus. This can be
done by a recursive pass through the tree. In the M-step, we must maximize the EM auxiliary function with respect to the parameters of the logistic regressions and multinomial
leaves as well as the input vectors ~x(w). The logistic regressions in the tree decouple and
can be optimized separately by Newton?s method, while the multinomial leaves have a simple closed-form update. Though the input vectors are shared across all logistic regressions
in the tree, we can compute their gradients and hessians in one recursive pass and update
them by Newton?s method as well.
The EM algorithm for HMEs converges to a local maximum of the log-likelihood, or equivalently, a local minimum of the training perplexity
?
?? C1
?Y
?
Ptrain =
Pr(wj |wi )Cij
,
(2)
?
?
ij
P
where C = ij Cij is the total number of observed bigrams in the training corpus. The
algorithm is sensitive to the choice of initialization; in particular, as we show in the next
word
initialized by
PCA or SDE
input
vector
logistic
regression
logistic
regression
logistic
regression
multinomial
distribution
multinomial
distribution
multinomial
distribution
multinomial
distribution
Figure 3: Two-layer HME for bigram modeling. Words are mapped to input vectors; probabilities of next words are computed by summing over paths through the tree. The mapping
from words to input vectors is initialized by dimensionality reduction of bigram counts.
Ptest
init
random
PCA
SDE
Ptest
m
8
16
32
64
d
4
468
406
385
8
407
364
361
12
378
362
360
16
373
351
355
Table 1: Test perplexities of HMEs
with different input dimensionalities
and initializations.
4
435
385
350
336
d
8
429
361
328
308
12
426
360
320
298
16
428
355
317
294
Table 2: Test perplexities of HMEs
with different input dimensionalities
and numbers of leaves.
section, initialization of the input vectors by PCA or SDE leads to significantly better models than random initialization. We initialized the logistic regressions in the HME to split
the input vectors recursively along their dimensions of greatest variance. The multinomial
distributions at leaf nodes were initialized by uniform distributions.
For an HME with m multinomial leaves and d-dimensional input vectors, the number of
parameters scales as O(V d + V m + dm). The resulting model can be therefore be much
more compact than a full bigram model over V words.
4
Results
We evaluated our models on the ARPA North American Business News (NAB) corpus.
Our training set contained 78 million words from a 60,000 word vocabulary. In the interest
of speed, we truncated the lowest-count bigrams from our training set. This left us with
a training set consisting of 1.7 million unique bigrams. The test set, untruncated, had 13
million words resulting in 2.1 million unique bigrams.
4.1
Empirical evaluation
Table 1 reports the test perplexities of several HMEs whose input vectors were initialized
in different ways. The number of mixture components (i.e., leaves of the HME) was fixed
at m = 16. In all cases, the inputs initialized by PCA and SDE significantly outperformed
random initialization. PCA and SDE initialization performed equally well for all but the
lowest-dimensional inputs. Here SDE outperformed PCA, most likely because the first few
eigenvectors of SDE capture more variance in the bigram counts than those of PCA (see
Figure 1).
Table 2 reports the test perplexities of several HMEs initialized by SDE, but with varying
input dimensionality (d) and numbers of leaves (m). Perplexity decreases with increasing
tree depth and input dimensionality, but increasing the dimensionality beyond d = 8 does
not appear to give much gain.
4.2
Comparison to a class-based bigram model
w
z
w'
Figure 4: Belief network for AMM.
We obtained baseline results from an AMM [13]
trained on the same corpus. The model (Figure 4)
has the form
X
Pr(w0 |w) =
Pr(z|w) ? Pr(w0 |z). (3)
z
The number of estimated parameters in AMMs
scales as 2?|Z|?V , where |Z| is the size of the latent variable (i.e., number of classes)
and V is the number of words in the vocabulary.
parameters (*1000)
960
1440
2400
4320
Ptest (AMM)
456
414
353
310
Ptest (HME)
429
361
328
308
improvement
6%
13%
7%
1%
Table 3: Test perplexities of HMEs and AMMs with roughly equal parameter counts.
Table 3 compares the test perplexities of several HMEs and AMMs with similar numbers
of parameters. All these HMEs had d = 8 inputs initialized by SDE. In all cases, the HMEs
match or outperform the AMMs. The performance is nearly equal for the larger models,
which may be explained by the fact that most of the parameters of the larger HMEs come
from the multinomial leaves, not from the distributed inputs.
4.3
Comparison to NPLM
The most successful large-scale application of distributed representations to language modeling is the NPLM of Bengio et al. [2, 3], which in part inspired our work. We now compare
the main aspects of the two models.
?
m
8
16
32
64
d
4
1
2
4
9
8
1
2
4
10
12
1
2
4
10
16
1
2
4
10
Table 4: Training times ? in hours
for HMEs with m leaves.
The NPLM uses softmax to compute the probability of a word w0 given its context, thus requiring a
separate normalization for each context. Estimating the parameters of this softmax requires O(V )
computation per observed context and accounts
for almost all of the computational resources required by the model. Because of this, the NPLM
vocabulary size was restricted to 18000 words,
and even then it required more than 3 weeks using 40 CPUs to finish 5 epochs of training [2].
By contrast, our HMEs require O(md) computation per observed bigram. As Table 4 shows, actual training times are rather insensitive to input dimensionality. This allowed us to use
a 3.5? larger vocabulary and a larger training corpus than were used for the NPLM, and
still complete training our largest models in a matter of hours. Note that the numbers in
Table 4 do not include the time to compute the initial distributed representations by PCA
(30 minutes) or SDE (3 days), but these computations do not need to be repeated for each
trained model.
The second difference between our model and the NPLM is the choice of initialization.
Bengio et al. [3] report negligible improvement from initializing the NPLM input vectors
by singular value decomposition. By contrast, we found that initialization by PCA or SDE
was essential for optimal performance of our models (Table 1).
Finally, the NPLM was applied to multiword contexts. We have not done these experiments yet, but our model extends naturally to multiword contexts, as we explain in the next
section.
5
Discussion
In this paper, we have presented a statistical language model that exploits hierarchical
distributed representations of word contexts. The model shares the advantages of the
NPLM [2], but differs in its use of dimensionality reduction for effective parameter ini-
tialization and in the significant speedup provided by the HME architecture. We can consequently scale our models to larger training corpora and vocabularies. We have also demonstrated that our models consistently match or outperform a baseline class-based bigram
model.
The class-based bigram model is nearly as effective as the HME, but it has the major drawback that there is no straightforward way to extend it to multiword contexts without exploding its parameter count. Like the NPLM, however, the HME can be easily extended.
We can form an input vector for a multiword history (w1 , w2 ) simply by concatenating the
vectors ~x(w1 ) and ~x(w2 ). The parameters of the corresponding HME can be learned by
an EM algorithm similar to the one in this paper. Initialization from dimensionality reduction is also straightforward: we can compute the low dimensional representation for each
word separately. We are actively pursuing these ideas to train models with hierarchical
distributed representations of multiword contexts.
References
[1] A. Y. Alfakih, A. Khandani, and H. Wolkowicz. Solving Euclidean distance matrix completion problems via semidefinite programming. Computational Optimization Applications, 12(13):13?30, 1999.
[2] Y. Bengio, R. Ducharme, P. Vincent, and C. Janvin. A neural probabilistic language model.
Journal of Machine Learning Research, 3:1137?1155, 2003.
[3] Y. Bengio, R. Ducharme, P. Vincent, and C. Jauvin. A neural probabilistic language model. In
T. K. Leen, T. G. Dietterich, and V. Tresp, editors, Advances in Neural Information Processing
Systems, volume 13, Cambridge, MA, 2001. MIT Press.
[4] D. B. Borchers. CSDP, a C library for semidefinite programming. Optimization Methods and
Software, 11(1):613?623, 1999.
[5] P. Brown, S. D. Pietra, V. D. Pietra, and R. Mercer. The mathematics of statistical machine
translation: parameter estimation. Computational Linguistics, 19(2):263?311, 1991.
[6] P. F. Brown, V. J. D. Pietra, P. V. deSouza, J. C. Lai, and R. L. Mercer. Class-based n-gram
models of natural language. Computational Linguistics, 18(4):467?479, 1992.
[7] S. Chen and J. Goodman. An empirical study of smoothing techniques for language modeling.
In Proceedings of the 34th Annual Meeting of the ACL, pages 310?318, 1996.
[8] M. Collins. Three generative, lexicalised models for statistical parsing. In Proceedings of the
35th Annual Meeting of the Association for Computational Linguistics, 1997.
[9] J. Ham, D. D. Lee, S. Mika, and B. Sch?olkopf. A kernel view of the dimensionality reduction of
manifolds. In Proceedings of the Twenty First International Conference on Machine Learning
(ICML-04), Banff, Canada, 2004.
[10] T. Hofmann and J. Puzicha. Statistical models for co-occurrence and histogram data. In Proceedings of the International Conference Pattern Recognition, pages 192?194, 1998.
[11] F. Jelinek. Statistical Methods for Speech Recognition. MIT Press, 1997.
[12] M. I. Jordan and R. A. Jacobs. Hierarchical mixtures of experts and the EM algorithm. Neural
Computation, 6:181?214, 1994.
[13] L. K. Saul and F. C. N. Pereira. Aggregate and mixed-order Markov models for statistical
language processing. In C. Cardie and R. Weischedel, editors, Proceedings of the Second Conference on Empirical Methods in Natural Language Processing (EMNLP-97), pages 81?89,
New Providence, RI, 1997.
[14] K. Q. Weinberger, F. Sha, and L. K. Saul. Learning a kernel matrix for nonlinear dimensionality
reduction. In Proceedings of the Twenty First International Confernence on Machine Learning
(ICML-04), Banff, Canada, 2004.
[15] C. Zhai and J. Lafferty. A study of smoothing methods for language models applied to information retrieval. ACM Transactions on Information Systems, 22(2):179?214, 2004.
| 2691 |@word polynomial:1 bigram:20 seek:1 covariance:2 decomposition:1 jacob:1 fifteen:1 recursively:1 reduction:15 initial:1 contains:2 yet:1 must:2 parsing:2 john:1 eleven:1 enables:1 hofmann:2 plot:1 interpretable:1 update:2 greedy:1 fewer:2 leaf:14 generative:1 record:1 gure:1 matrix1:1 node:2 banff:2 monday:1 five:1 along:2 saturday:1 combine:1 introduce:1 pairwise:2 upenn:1 indeed:1 roughly:1 themselves:1 frequently:1 inspired:2 cpu:1 enumeration:1 window:4 solver:2 increasing:2 actual:1 provided:1 estimating:1 maximizes:1 mass:1 lowest:2 sde:17 developed:2 multidimensional:2 prohibitively:1 appear:1 negligible:1 walnut:1 understood:1 local:4 despite:1 encoding:1 path:4 interpolation:1 might:1 acl:1 mika:1 initialization:12 studied:1 co:2 statistically:1 unique:2 practice:2 recursive:2 definite:1 differs:2 nite:5 empirical:3 significantly:4 projection:2 word:54 cannot:2 close:1 onto:1 interior:1 context:36 map:2 demonstrated:1 maximizing:2 straightforward:2 convex:2 lexicalised:1 spanned:1 embedding:10 analogous:1 play:1 massive:1 programming:6 us:1 origin:1 pa:1 element:4 infrequently:1 recognition:3 expensive:2 observed:6 levine:1 role:1 bottom:1 solved:1 capture:3 parameterize:1 thousand:1 initializing:1 wj:2 connected:1 kilian:1 news:1 decrease:1 ham:1 ideally:1 multiword:9 trained:3 solving:1 predictive:2 easily:1 represented:2 train:2 effective:3 borchers:1 couldn:1 aggregate:2 neighborhood:3 whose:2 larger:7 ducharme:2 unseen:2 nab:1 semide:5 advantage:3 eigenvalue:6 interaction:1 product:1 frequent:3 neighboring:2 mixing:1 olkopf:1 exploiting:1 rst:2 regularity:4 billion:1 incremental:2 converges:1 blitzer:2 completion:1 fixing:1 nearest:1 ij:4 expectationmaximization:1 auxiliary:1 predicted:4 involves:1 kilianw:1 come:1 concentrate:1 drawback:1 centered:1 human:1 broke:2 require:1 assign:1 suffices:1 alleviate:1 extension:1 hall:1 lawrence:1 mapping:4 week:1 major:1 adopt:1 estimation:1 outperformed:2 sensitive:1 largest:2 leftover:1 mit:2 sunday:1 rather:1 varying:1 derived:3 june:1 improvement:5 consistently:1 likelihood:2 contrast:3 baseline:2 jauvin:1 typically:1 hmes:14 lsaul:1 semantics:1 djk:1 among:1 art:1 smoothing:4 special:1 initialize:1 softmax:2 equal:2 unsupervised:2 nearly:2 icml:2 semipositive:1 report:3 few:1 simultaneously:1 preserve:3 individual:1 pietra:3 consisting:1 freedom:1 interest:1 evaluation:1 mixture:9 sixty:1 semidefinite:4 accurate:2 necessary:1 tree:10 euclidean:4 initialized:10 amm:6 arpa:1 modeling:9 expects:1 rare:1 uniform:1 csdp:1 successful:1 dij:6 dependency:1 providence:1 twelve:1 international:3 probabilistic:5 off:1 lee:1 together:1 w1:2 squared:5 central:1 opposed:1 choose:1 emnlp:1 expert:9 american:1 leading:1 friday:1 actively:1 account:1 hme:16 north:1 matter:1 performed:2 root:1 view:1 closed:1 nplm:15 minimize:1 variance:9 generalize:1 vincent:2 cardie:1 multiplying:1 researcher:1 history:1 explain:1 sharing:1 frequency:7 dm:1 naturally:1 associated:1 gain:1 seventeen:1 wolkowicz:1 knowledge:1 ptrain:1 dimensionality:22 khandani:1 back:1 cik:1 dt:1 day:1 april:1 leen:1 done:3 though:3 evaluated:1 generality:1 furthermore:1 ptest:4 nonlinear:13 logistic:8 grows:1 building:1 effect:1 dietterich:1 normalized:2 requiring:1 brown:2 symmetric:1 semantic:1 won:1 criterion:2 ini:1 occuring:1 demonstrate:1 complete:1 syntactically:1 image:1 recently:2 common:1 rotation:1 multinomial:12 fourteen:1 exponentially:1 insensitive:1 million:6 discussed:1 extend:1 volume:1 association:1 significant:3 refer:1 cambridge:1 automatic:1 mathematics:1 language:23 had:2 dot:1 longer:1 posterior:1 apart:2 perplexity:10 binary:1 meeting:2 captured:3 minimum:1 preceding:1 fernando:1 maximize:3 july:1 ii:2 branch:1 full:2 exploding:1 faster:1 match:2 retrieval:2 lai:1 equally:1 involving:1 regression:8 multilayer:2 metric:1 histogram:1 represent:1 normalization:4 kernel:2 achieved:3 c1:1 preserved:1 separately:2 singular:1 crucial:1 goodman:1 container:2 meaningless:1 w2:2 sch:1 subject:2 december:1 lafferty:1 jordan:1 bengio:6 iii:1 split:1 variety:1 xj:3 finish:1 weischedel:1 pennsylvania:1 architecture:3 competing:1 idea:1 six:1 pca:19 speech:2 hessian:1 nine:1 generally:1 eigenvectors:2 discount:1 ten:1 concentrated:1 simplest:2 specifies:2 outperform:2 judging:1 estimated:1 per:4 discrete:1 express:1 four:2 pj:1 v1:1 vast:1 graph:1 sum:2 angle:1 extends:1 almost:1 pursuing:1 scaling:1 comparable:1 bit:1 capturing:1 layer:2 followed:1 annual:2 occur:2 constraint:1 ri:1 software:1 encodes:1 aspect:2 speed:1 extremely:1 speedup:1 department:1 structured:2 combination:1 march:1 across:3 remain:1 slightly:1 ascertain:1 em:5 wi:3 making:2 untruncated:1 intuitively:1 explained:1 pr:8 restricted:1 resource:1 discus:1 count:9 fail:1 letting:1 fed:2 end:1 apply:1 eight:1 hierarchical:9 spectral:1 occurrence:1 weinberger:4 original:1 top:2 remaining:1 include:1 linguistics:3 newton:2 exploit:2 february:1 already:1 costly:1 sha:1 md:1 september:1 gradient:1 subspace:3 distance:12 separate:3 mapped:1 street:1 majority:1 w0:9 seven:1 manifold:1 trivial:1 water:2 assuming:1 length:1 relationship:2 zhai:1 equivalently:1 difficult:2 unfortunately:1 cij:3 thirteen:1 october:1 stated:1 proper:1 twenty:2 perform:1 markov:2 eighteen:1 enabling:1 november:1 january:1 truncated:1 extended:1 locate:1 discovered:1 smoothed:1 august:1 canada:2 required:2 janvin:1 sentence:3 optimized:1 learned:1 hour:2 beyond:1 suggested:1 pattern:1 program:1 belief:1 greatest:1 event:3 natural:5 rely:1 difficulty:1 business:1 vase:4 representing:1 scheme:1 library:1 axis:1 philadelphia:1 tresp:1 text:1 epoch:1 tantamount:1 relative:1 loss:1 expect:1 fully:1 mixed:1 sixteen:1 pij:3 consistent:1 sufficient:1 mercer:2 editor:2 storing:1 pi:1 balancing:1 translation:3 share:1 elsewhere:1 truncation:1 saul:4 neighbor:1 jelinek:1 distributed:13 dimension:7 vocabulary:12 valid:1 gram:6 avoids:1 rich:1 computes:2 depth:1 far:2 transaction:1 compact:1 global:1 desouza:1 reveals:1 corpus:16 summing:1 xi:7 spectrum:2 continuous:2 latent:3 table:13 learn:1 init:1 tuesday:1 complex:2 main:3 allowed:1 repeated:1 fig:2 depicts:1 pereira:4 concatenating:1 lie:2 formula:1 unwieldy:1 embed:2 minute:1 showing:1 essential:2 ci:1 magnitude:1 conditioned:1 occurring:1 sparseness:1 push:1 chen:1 suited:1 simply:1 likely:1 contained:1 wednesday:1 acm:1 ma:1 conditional:4 goal:3 formulated:1 consequently:2 shared:1 change:1 semantically:9 decouple:1 principal:2 total:1 gij:1 pas:2 meaningful:1 thursday:1 puzicha:2 collins:1 dissimilar:2 evaluate:1 |
1,863 | 2,692 | Limits of Spectral Clustering
Ulrike von Luxburg and Olivier Bousquet
Max Planck Institute for Biological Cybernetics
Spemannstr. 38, 72076 T?ubingen, Germany
{ulrike.luxburg,olivier.bousquet}@tuebingen.mpg.de
Mikhail Belkin
The University of Chicago, Department of Computer Science
1100 E 58th st., Chicago, USA
[email protected]
Abstract
An important aspect of clustering algorithms is whether the partitions
constructed on finite samples converge to a useful clustering of the whole
data space as the sample size increases. This paper investigates this
question for normalized and unnormalized versions of the popular spectral clustering algorithm. Surprisingly, the convergence of unnormalized
spectral clustering is more difficult to handle than the normalized case.
Even though recently some first results on the convergence of normalized spectral clustering have been obtained, for the unnormalized case
we have to develop a completely new approach combining tools from
numerical integration, spectral and perturbation theory, and probability.
It turns out that while in the normalized case, spectral clustering usually
converges to a nice partition of the data space, in the unnormalized case
the same only holds under strong additional assumptions which are not
always satisfied. We conclude that our analysis gives strong evidence for
the superiority of normalized spectral clustering. It also provides a basis
for future exploration of other Laplacian-based methods.
1
Introduction
Clustering algorithms partition a given data set into several groups based on some notion
of similarity between objects. The problem of clustering is inherently difficult and often
lacks clear criteria of ?goodness?. Despite the difficulties in determining the quality of a
given partition, it is still possible to study desirable properties of clustering algorithms from
a theoretical point of view. In this paper we study the consistency of spectral clustering,
which is an important property in the general framework of statistical pattern recognition.
A clustering algorithm is consistent if it produces a well-defined (and, hopefully, sensible)
partition, given sufficiently many data points. The consistency is a basic sanity check, as an
algorithm which is not consistent would change the partition indefinitely as we add points to
the dataset, and, consequently, no reasonable small-sample performance could be expected
at all. Surprisingly, relatively little research into consistency of clustering algorithms has
been done so far, exceptions being only k-centers (Pollard, 1981) and linkage algorithms
(Hartigan, 1985).
While finite-sample properties of spectral clustering have been studied from a theoretical
point of view (Spielman and Teng, 1996; Guattery and Miller, 1998; Kannan et al., 2000;
Ng et al., 2001; Meila and Shi, 2001) we focus on the limit behavior for sample size tending
to infinity. In this paper we develop a new strategy to prove convergence results for spectral
clustering algorithms. Unlike our previous attempts this strategy allows to obtain results
for both normalized and unnormalized spectral clustering. As a first result we can recover
the main theorem of von Luxburg et al. (2004), which had been proved with different and
more restrictive methods, and, in brief, states that usually normalized spectral clustering
converges. We also extend that result to the case of multiple eigenvectors. Our second result
concerns the case of unnormalized spectral clustering, for which no convergence properties
had been known so far. This case is much more difficult to treat than the normalized
case, as the limit operators have a more complicated form. We show that unnormalized
spectral clustering also converges, but only under strong additional assumptions. Contrary
to the normalized case, those assumptions are not always satisfied, as we can show by
constructing an example, and in this case there is no hope for convergence. Even worse,
on a finite sample it is impossible to verify whether the assumptions hold or not. As a third
result we prove statements about the form of the limit clustering. It turns out that in case
of convergence, the structure of the clustering constructed on finite samples is conserved in
the limit process. From this we can conclude that if convergence takes place, then the limit
clustering presents an intuitively appealing partition of the data space.
It is also interesting to note that several recent methods for semi-supervised and transductive learning are based on eigenvectors of similarity graphs (cf. Belkin and Niyogi, 2003;
Chapelle et al., 2003; Zhu et al., 2003). Our theoretical framework can also be applied to
investigate the consistency of those algorithms with respect to the unlabeled data.
There is an ongoing debate on the advantages of the normalized versus unnormalized graph
Laplacians for spectral clustering. It has been found empirically that the normalized version
performs as well or better than the unnormalized version (e.g., Van Driessche and Roose,
1995; Weiss, 1999; in the context of semi-supervised learning see also Zhou et al., 2004).
We are now able to provide additional evidence to this effect from a theoretical point of
view. Normalized spectral clustering is a well-behaved algorithm which always converges
to a sensible limit clustering. Unnormalized spectral clustering on the other hand should be
treated with care as consistency can only be asserted under strong assumptions which are
not always satisfied and, moreover, are difficult to check in practice.
2
Graph Laplacians and spectral clustering on finite samples
In the following we denote by ?(T ) the spectrum of a linear operator, by C(X ) the space
of continuous functions on X with infinity norm, and by rg(d) the range of a function
d ? C(X ). For given sample points X1 , ..., Xn drawn iid according to an (unknown) distribution P on some data space X we denote the empirical distribution by Pn . For a nonnegative, symmetric similarity function s :PX ?X ? IR we define the similarity matrix
n
as Kn := (s(Xi , Xj ))i,j=1,...,n , set di := j=1 s(Xi , Xj ), and define the degree matrix
Dn as the diagonal matrix with entries di . The unnormalized Laplacian matrix is defined
?1/2
?1/2
as Ln := Dn ? Kn , and two common ways of normalizing it are L?n := Dn Ln Dn
??
?1
or Ln := Dn Ln . In the following we always arrange the eigenvalues of the Laplacian
matrices in non-decreasing order 0 = ?1 ? ?2 ... ? ?n respecting their multiplicities. In its
simplest form, unnormalized (resp. normalized) spectral clustering partitions the sample
points Xi into two groups according to whether the i-th coordinate of the second eigenvector is larger or smaller than a certain threshold b ? IR. Often, instead of considering
only the second eigenvector, one uses the first r eigenvectors (for some small number r)
simultaneously to obtain a partition into several sets. For an overview of different spectral
clustering algorithms see for example Weiss (1999).
3
Limit results
In this section we want to state and discuss our main results. The general assumptions
in the following three theorems are that the data space X is a compact metric space from
which the sample points (Xi )i?IN are drawn independently according to an unknown probability distribution P . Moreover we require the similarity function s : X ?X ? IR to be
continuous, and in the normalized case to be bounded away from 0, that is s(x, y) > l > 0
for all x, y ? X and some l ? IR. By d ? C(X ) we will denote the ?degree function?, and
U ? and U will denote the ?limit operators? of L?n and Ln for n ? ?. The exact definitions
of these functions and operators, as well as all further mathematical details, definitions, and
proofs will be postponed to Section 4.
Let us start with the first question raised in the introduction: does the spectral clustering
constructed on a finite sample converge to a partition of the whole data space if the sample
size increases? In the normalized case, convergence results have recently been obtained
in von Luxburg et al. (2004). However, those methods were specifically designed for the
normalized Laplacian and cannot be used in the unnormalized case. Here we state a convergence result for the normalized case in the form how it can be obtained with our new
methods. The theorem is formulated for the symmetric normalization L?n , but it holds
similarly for the normalization L??n .
Theorem 1 (Convergence of normalized spectral clustering) Under the general assumptions, if the first r eigenvalues of the limit operator U ? have multiplicity 1, then the
same holds for the first r eigenvalues of L?n for sufficiently large n. In this case, the first r
eigenvalues of L?n converge to the first r eigenvalues of U ? , and the corresponding eigenvectors converge almost surely. The partitions constructed by normalized spectral clustering
from the first r eigenvectors on finite samples converge almost surely to a limit partition of
the whole data space.
Our new result about the convergence in the unnormalized case is the following:
Theorem 2 (Convergence of unnormalized spectral clustering) Under the general assumptions, if the first r eigenvalues of the limit operator U have multiplicity 1 and are not
element of rg(d), then the same holds for the first r eigenvalues of n1 Ln for sufficiently
large n. In this case, the first r eigenvalues of n1 Ln converge to the first r eigenvalues of U ,
and the the corresponding eigenvectors converge almost surely. The partitions constructed
by unnormalized spectral clustering from the first r eigenvectors on finite samples converge
almost surely to a limit partition of the whole data space.
On the first glance, this theorem looks very similar to Theorem 1: if the general assumptions are satisfied and the first eigenvalues are ?nice?, then unnormalized spectral clustering converges. However, the difference between Theorems 1 and 2 is what it means for
an eigenvalue to be ?nice?. In Theorem 1 we only require the eigenvalues to have multiplicity 1 (and in fact, if the multiplicity is larger than 1 we can still prove convergence
of eigenspaces instead of eigenvectors). In Theorem 2, however, the condition ? 6? rg(d)
has to be satisfied. In the proof this is needed to ensure that the eigenvalue ? is isolated
in the spectrum of U , which is a fundamental requirement for applying perturbation theory
to the convergence of eigenvectors. If this condition is not satisfied, perturbation theory
is in principle unsuitable to obtain convergence results for eigenvectors. The reason why
this condition appears in the unnormalized case but not in the normalized case lies in the
structure of the respective limit operators, which, surprisingly, is more complicated in the
unnormalized case than in the normalized one. In the next section we will construct an example where the second eigenvalue indeed lies within rg(d). This means that there actually
exist situations in which Theorem 2 cannot be applied, and hence unnormalized spectral
clustering might not converge.
Now we want to turn to the second question raised in the introduction: In case of convergence, is the limit clustering a reasonable clustering of the whole data space? To answer
this question we analyze the structure of the limit operators (for simplicity we state this
for the unnormalized case only). Assume that we are given a partition X = ?ki=1 Xi
of the data space into k disjoint sets. If we order the sample points according to their
memberships in the sets Xi , then we can write the Laplacian in form of a block matrix
Ln ? (Lij,n )i,j=1,...,k where each sub-matrix Lij,n contains the rows of Ln corresponding
to points in set Xi and the columns corresponding to points in Xj . In a similar way, the limit
operator U can be decomposed into a matrix of operators Uij : C(Xj ) ? C(Xi ). Now we
can show that for all i, j = 1, ..., k the sub-matrices n1 Lij,n converge to the corresponding
sub-operators Uij such that their spectra converge in the same way as in Theorems 1 and 2.
This is a very strong result as it means that for every given partition of X , the structure of
the operators is preserved in the limit process.
Theorem 3 (Structure of the limit operators) Let X = ?ki=1 Xi be a partition of the data
space. Let Lij,n be the sub-matrices of Ln introduced above, Uij : C(Xj ) ? C(Xi ) the
?
?
restrictions of U corresponding to the sets Xi and Xj , and Uij,n
and Uij
the analogous
1
quantities for the normalized case. Then under the general assumptions, n Lij,n converges
?
compactly to Uij a.s. and L?ij,n converges compactly to Uij
a.s.
With this result it is then possible to give a first answer on the question how the limit
partitions look like. In Meila and Shi (2001) it has been established that normalized spectral
clustering tries to find a partition such that a random walk on the sample points tends to
stay within each of the partition sets Xi instead of jumping between them. With the help
of Theorem 3, the same can now be said for the normalized limit partition, and this can
also be extended to the unnormalized case. The operators U ? and U can be interpreted as
diffusion operators on the data space. The limit clusterings try to find a partition such that
the diffusion tends to stay within the sets Xi instead of jumping between them. In particular,
the limit partition segments the data space into sets such that the similarity within the sets
is high and the similarity between the sets is low, which intuitively is what clustering is
supposed to do.
4
Mathematical details
In this section we want to explain the general constructions and steps that need to be
taken to prove Theorems 1, 2, and 3. However, as the proofs are rather technical we only
present proof sketches that convey the overall strategy. Detailed proofs can be found in von
Luxburg (2004) where all proofs are spelled out in full length. Moreover, we will focus on
the proof of Theorem 2 as the other results can be proved similarly.
To be able to define convergence of linear operators, all operators have to act on the same
space. As this is not the case for the matrices Ln for different n, for each Ln we will
construct a related operator Un on the space C(X ) which will be used instead of Ln . In
Step 2 we show that the interesting eigenvalues and eigenvectors of n1 Ln and Un are in a
one-to-one relationship. Then we will prove that the Un converge in a strong sense to some
limit operator U on C(X ) in Step 3. As we can show in Step 4, this convergence implies
the convergence of eigenvalues and eigenvectors of Un . Finally, assembling the parts will
finish the proof of Theorem 2.
Step 1: Construction of the operators Un on C(X ).
We first define the empirical and true degree functions in C(X ) as
Z
Z
dn (x) := s(x, y)dPn (y)
and
d(x) := s(x, y)dP (y).
Corresponding to the matrices Dn and Kn we introduce the following multiplication and
integral operators on C(X ):
Mdn f (x) := dn (x)f (x)
Z
Sn f (x) := s(x, y)f (y)dPn (y)
and
and
Md f (x) := d(x)f (x)
Z
Sf (x) := s(x, y)f (y)dP (y).
Note that dn (Xi ) = n1 di , and for f ? C(X ) and v := (f (X1 ), ..., f (Xn ))? it holds
that n1 (Dn v)i = Mdn f (Xi ) and n1 (Kn v)i = Sn f (Xi ). Hence the function dn and the
operators Mdn and Sn are the counterparts of the discrete degrees n1 di and the matrices
1
1
n Dn and n Kn . The scaling factor 1/n comes from the hidden 1/n-factor in the empirical
distribution Pn . The natural pointwise limits of dn , Mdn , and Sn for n ? ? are given
by d, Md , and S. The operators corresponding to the unnormalized Laplacians n1 Ln =
1
n (Dn ? Kn ) and its limit operator are
Un f (x) := Mdn f (x) ? Sn f (x)
and
U f (x) := Md f (x) ? Sf (x).
1
Step 2: Relations between ?( n
Ln ) and ?(Un ).
Proposition 4 (Spectral properties) 1. The spectrum of Un consists of rg(dn ), plus some
isolated eigenvalues with finite multiplicity. The same holds for U and rg(d).
2. If f ? C(X ) is an eigenfunction of Un with arbitrary eigenvalue ?, then the vector
v ? IRn with vi = f (Xi ) is an eigenvector of the matrix n1 Ln with eigenvalue ?.
3. If v is an eigenvector of the matrix n1 Ln with eigenvalue ? 6? rg(dn ), then the
P
function f (x) = n1 ( j s(x, Xj )vj )/(dn (x) ? ?) is the unique eigenfunction of Un with
eigenvalue ? satisfying f (Xi ) = vi .
Proof. It is well-known that the (essential) spectrum of a multiplication operator coincides
with the range of the multiplier function. Moreover, the spectrum of a sum of a bounded
operator with a compact operator contains the essential spectrum of the bounded operator.
Additionally, it can only contain some isolated eigenvalues with finite multiplicity (e.g.,
Theorem IV.5.35 in Kato, 1966). The proofs of the other parts of this proposition can be
obtained by elementary shuffling of eigenvalue equations and will be skipped.
,
Step 3: Convergence of Un to U .
Dealing with the randomness. Recall that the operators Un are random operators as
they depend on the given sample points X1 , ..., Xn via the empirical distribution Pn . One
important tool to cope with this randomness will be the following proposition:
Proposition 5 (Glivenko-Cantelli class) Let (X , d) be a compact metric space and s :
X ?X ? RIR continuous. Then
R F := {s(x, ?); x ? X } is a Glivenko-Cantelli class, that is
supx?X | s(x, y)dPn (y) ? s(x, y)dP (y)| ? 0 almost surely.
Proof. This proposition follows from Theorem 2.4.1. of v. d. Vaart and Wellner (1996). ,
Note that one direct consequence of this proposition is that kdn ? dk? ? 0 a.s.
Types of convergence. Let E be an arbitrary Banach space and B its unit
S ball. A sequence
(Sn )n of linear operators on E is called collectively compact if the set n Sn B is relatively
compact in E (with respect to the norm topology). A sequence of operators converges
collectively compactly if it converges pointwise and if there exists some N ? IN such that
the operators (Sn ? S)n>N are collectively compact. A sequence of operators converges
compactly if it converges pointwise and if for every sequence (xn )n in B, the sequence
(S?Sn )xn is relatively compact. See Anselone (1971) and Chatelin (1983) for background
reading. A sequence (xn )n in E converges up to a change of sign to x ? E if there exists
a sequence (an )n of signs an ? {?1, +1} such that the sequence (an xn )n converges to x.
Proposition 6 (Un converges compactly to U a.s.) Let X be a compact metric space and
s : X ?X ? IR continuous. Then Un converges to U compactly a.s.
Proof. (a) Sn converges to S collectively compactly a.s. With the help of the GlivenkoCantelli property in Proposition 5 it is easy to see that Sn converges to S pointwise, that
is kSn f ? Sf k? ? 0 a.s. for all f ? C(X ). As the limit operator S is compact, to
prove that (Sn ? S)n are collectively compact a.s. it is enough to prove that (Sn )n are
collectively compact a.s. This can be done by the Arzela-Ascoli theorem.
(b) Mdn converges to Md in operator norm a.s. This is a direct consequence of the
Glivenko-Cantelli properties of Proposition 5.
(c) Un = Sn ? Mdn converges to U = S ? Md compactly a.s. Both operator
norm convergence and collectively compact convergence imply compact convergence
(cf. Proposition 3.18 of Chatelin, 1983). Moreover, it is easy to see that the sum of two
compactly converging operators converges compactly.
,
Step 4: Convergence of the eigenfunctions of Un to those of U .
It is a result of perturbation theory (see the comprehensive treatment in Chatelin, 1983,
especially Section 5.1) that compact convergence of operators implies the convergence of
eigenvalues and spectral projections in the following way. If ? is an isolated eigenvalue in
?(U ) with finite multiplicity, then there exists a sequence ?n ? ?(Un ) of isolated eigenvalues with finite multiplicity such that ?n ? ?. If the first r eigenvalues of T have
multiplicity 1, then the same holds for the first r eigenvalues of Tn for sufficiently large
n, and the i-th eigenvalues of Tn converge to the i-th eigenvalue of T . The corresponding
eigenvectors converge up to a change of sign. If the multiplicity of the eigenvalues is larger
than 1 but finite, then the corresponding eigenspaces converge. Note that for eigenvalues
which are not isolated in the spectrum, convergence cannot be asserted, and the same holds
for the corresponding eigenvectors (e.g., Section IV.3 of Kato, 1966).
In our case, by Proposition 4 we know that the spectrum of U consists of the whole interval rg(d), plus eventually some isolated eigenvalues. Hence an eigenvalue ? ? ?(U )
is isolated in the spectrum iff ? 6? rg(d) holds, in which case convergence holds as stated
above.
Step 5: Convergence of unnormalized spectral clustering.
Now we can to put together the different parts. In the first two steps we transferred the
problem of the convergence of the eigenvectors of n1 Ln to the convergence of eigenfunctions of Un . In Step 3 we showed that Un converges compactly to the limit operator U ,
which according to Step 4 implies the convergence of the eigenfunctions of Un . In terms
of the eigenvectors of n1 Ln this means the following: if ? denotes the j-th eigenvalue
of U with eigenfunction f ? C(X ) and ?n the j-th eigenvalue of n1 Ln with eigenvector vn = (vn,1 , ..., vn,n )? , then there exists a sequence of signs ai ? {?1, +1} such that
supi=1,...,n |ai vn,i ? f (Xi )| ? 0 a.s. As spectral clustering is constructed from the coordinates of the eigenvectors, this leads to the convergence of spectral clustering in the
unnormalized case. This completes the proof of Theorem 2.
,
The proof for Theorem 1 can be obtained in a very similar way. Here the limit operator is
Z
p
U ? f (x) := (I ? S ? )f (x) := f (x) ? (s(x, y)/ d(x)d(y) )f (y)dP (y).
The main difference to the unnormalized case is that the operator Md in U gets replaced by
the identity operator I in U ? . This simplifies matters as one can easily express the spectrum
of (I ? S ? ) via the spectrum of the compact operator S ? . From a different point of view,
consider the identity operator as the operator of multiplication by the constant one function
1. Its range is the single point rg(1) = {1}, and hence the critical interval rg(d) ? ?(U )
shrinks to the point 1 ? ?(U ? ), which in general is a non-isolated eigenvalue with infinite
multiplicity.
Finally, note that it is also possible to prove more general versions of Theorems 1 and 2
where the eigenvalues have finite multiplicity larger than 1. Instead of the convergence of
the eigenvectors we then obtain the convergence of the projections on the eigenspaces.
The proof of Theorem 3 works as the ones of the other two theorems. The exact definitions
of the operators considered in this case are
Z
q
?
Uij : C(Xj ) ? C(Xi ), ?ij fi (x) ? (sij (x, y)/ di (x)dj (y) )fj (y)dPj (y)
Z
Uij : C(Xj ) ? C(Xi ), Uij f (x) = ?ij di (x)fi (x) ? sij (x, y)fj (y)dPj (y)
where di , fi , Pi , and sij denote the restrictions of the functions to Xi and Xi ?Xj , respectively, and ?ij is 1 if i = j and 0 otherwise. For the diffusion interpretation, note that if
there exists an ideal partition of the data space (that is, s(xi , xj ) = 0 for xi , xj in different
?
sets Xi and Xj ), then the off-diagonal operators Uij
and Uij with i 6= j vanish, and the
?
first k eigenvectors of U and U can be reconstructed by the piecewise constant eigenvectors of the diagonal operators Uii? and Uii . In this situation, spectral clustering recovers
the ideal clustering. If there exists no ideal clustering, but there exists a partition such that
the off-diagonal operators are ?small? and the diagonal operators are ?large?, then it can
be seen by perturbation theory arguments that spectral clustering will find such a partition.
The off-diagonal operators can be interpreted as diffusion operators between different clusters (note that even in the unnormalized case, the multiplication operator only appears in
the diagonal operators). Hence, constructing a clustering with small off-diagonal operators
corresponds to a partition such that few diffusion between the clusters takes place.
Finally, we want to construct an example where the second eigenvalue of U satisfies ? ?
rg(d). Let X = [1, 2] ? IR, s(x, y) := xy, and p a piecewise constant probability density
on X with p(x) = c if 4/3 ? x < 5/3 and p(x) = (3 ? c)/2 otherwise, for some fixed
constant c ? [0, 3] (e.g., for small c this density has two clearly separated high density
regions). The degree function in this case is d(x) = 1.5x (independently of c) and has
range [1.5, 3] on X . We can see that an eigenfunction of U Rfor eigenvalue ? 6? rg(d) has
the form f (x) = ?x/(3x ? ?), where the equation ? = x2 /(3x ? ?)p(x)dx has to
be satisfied. This means that ? 6? rg(d) is an eigenvalue of U iff the equation g(?) :=
R2 2
!
x /(3x ? ?)p(x)dx = 1 is satisfied. For our simple density function p, this integral
1
can be solved analytically. It can then been seen that g(?) = 1 is only satisfied for ? = 0,
hence the only eigenvalue outside of rg(d) is the trivial eigenvalue 0.
Note that in applications of spectral clustering, we do not know the limit operator U and
hence cannot test whether its relevant eigenvalues are in its essential spectrum rg(d) or
not. If, for some special reason, one really wants to use unnormalized spectral clustering,
one should at least estimate the critical region rg(d) by [mini di /n, maxi di /n] and check
whether the relevant eigenvalues of n1 Ln are inside or close to this interval or not. This observation then gives an indication whether the results obtained can considered to be reliable
or not. However, this observation is not a valid statistical test.
5
Conclusions
We have shown that under standard assumptions, normalized spectral clustering always
converges to a limit partition of the whole data space which depends only on the probability
distribution P and the similarity function s. For unnormalized spectral clustering, this can
only be guaranteed under the strong additional assumption that the first eigenvalues of the
Laplacian do not fall inside the range of the degree function. As shown by our example,
this condition has to be taken seriously.
Consistency results are a basic sanity check for behavior of statistical learning algorithms.
Algorithms which do not converge cannot be expected to exhibit reliable results on finite
samples. Therefore, in the light of our theoretical analysis we assert that the normalized
version of spectral clustering should be preferred in practice. This suggestion also extends
to other applications of graph Laplacians including semi-supervised learning.
References
P. Anselone. Collectively compact operator approximation theory. Prentice-Hall, 1971.
M. Belkin and P. Niyogi. Using manifold structure for partially labeled classification. In
Advances in Neural Information Processing Systems 15, 2003.
O. Chapelle, J. Weston, and B. Sch?olkopf. Cluster kernels for semi-supervised learning. In
Advances in Neural Information Processing Systems 15, 2003.
F. Chatelin. Spectral Approximation of Linear Operators. Academic Press, 1983.
S. Guattery and G. L. Miller. On the quality of spectral separators. SIAM Journal of Matrix
Anal. Appl., 19(3), 1998.
J. Hartigan. Statistical theory in clustering. Journal of classification, 2:63?76, 1985.
R. Kannan, S. Vempala, and A. Vetta. On clusterings - good, bad and spectral. Technical
report, Computer Science Department, Yale University, 2000.
T. Kato. Perturbation theory for linear operators. Springer, Berlin, 1966.
M. Meila and J. Shi. A random walks view of spectral segmentation. In 8th International
Workshop on Artificial Intelligence and Statistics, 2001.
A. Ng, M. Jordan, and Y. Weiss. On spectral clustering: Analysis and an algorithm. In
Advances in Neural Information Processing Systems 14, 2001.
D. Pollard. Strong consistency of k-means clustering. Ann. of Stat., 9(1):135?140, 1981.
D. Spielman and S. Teng. Spectral partitioning works: planar graphs and finite element
meshes. In 37th Annual Symposium on Foundations of Computer Science, 1996.
A. v. d. Vaart and J. Wellner. Weak Convergence and Empirical Processes. Springer, 1996.
R. Van Driessche and D. Roose. An improved spectral bisection algorithm and its application to dynamic load balancing. Parallel Comput., 21(1), 1995.
U. von Luxburg. Statistical Learning with Similarity and Dissimilarity Functions. PhD
thesis, draft, available at http://www.kyb.tuebingen.mpg.de/?ule, 2004.
U. von Luxburg, O. Bousquet, and M. Belkin. On the convergence of spectral clustering
on random samples: the normalized case. In COLT, 2004.
Y. Weiss. Segmentation using eigenvectors: A unifying view. In Proceedings of the International Conference on Computer Vision, pages 975?982, 1999.
D. Zhou, O. Bousquet, T. Lal, J. Weston, and B. Sch?olkopf. Learning with local and global
consistency. In Advances in Neural Information Processing Systems 16, 2004.
X. Zhu, Z. Ghahramani, and J. Lafferty. Semi-supervised learning using Gaussian fields
and harmonic functions. In ICML, 2003.
| 2692 |@word version:5 norm:4 contains:2 seriously:1 dpn:3 dx:2 mesh:1 numerical:1 chicago:2 partition:28 kyb:1 designed:1 intelligence:1 indefinitely:1 provides:1 draft:1 mathematical:2 dn:17 constructed:6 direct:2 symposium:1 prove:8 consists:2 inside:2 introduce:1 indeed:1 expected:2 behavior:2 mpg:2 decreasing:1 decomposed:1 little:1 considering:1 moreover:5 bounded:3 what:2 interpreted:2 eigenvector:5 assert:1 every:2 act:1 partitioning:1 unit:1 superiority:1 planck:1 local:1 treat:1 tends:2 limit:31 consequence:2 anselone:2 despite:1 might:1 plus:2 studied:1 appl:1 range:5 unique:1 practice:2 block:1 empirical:5 projection:2 get:1 cannot:5 unlabeled:1 close:1 operator:60 put:1 context:1 impossible:1 applying:1 prentice:1 restriction:2 www:1 center:1 shi:3 independently:2 simplicity:1 handle:1 notion:1 coordinate:2 analogous:1 resp:1 construction:2 exact:2 olivier:2 us:1 element:2 recognition:1 satisfying:1 labeled:1 solved:1 region:2 respecting:1 dynamic:1 depend:1 segment:1 completely:1 basis:1 compactly:11 easily:1 separated:1 glivenko:3 artificial:1 outside:1 sanity:2 larger:4 otherwise:2 niyogi:2 statistic:1 vaart:2 transductive:1 advantage:1 eigenvalue:46 sequence:10 indication:1 relevant:2 combining:1 kato:3 iff:2 supposed:1 olkopf:2 convergence:38 cluster:3 requirement:1 produce:1 converges:22 spelled:1 object:1 help:2 develop:2 stat:1 ij:4 strong:8 c:1 implies:3 come:1 rfor:1 exploration:1 require:2 really:1 proposition:11 biological:1 elementary:1 hold:11 sufficiently:4 considered:2 hall:1 arrange:1 tool:2 hope:1 clearly:1 always:6 gaussian:1 rather:1 zhou:2 pn:3 focus:2 check:4 cantelli:3 skipped:1 sense:1 membership:1 hidden:1 uij:12 relation:1 irn:1 germany:1 overall:1 classification:2 colt:1 raised:2 integration:1 special:1 field:1 construct:3 ng:2 look:2 icml:1 future:1 report:1 piecewise:2 belkin:4 few:1 simultaneously:1 comprehensive:1 replaced:1 n1:16 attempt:1 investigate:1 misha:1 light:1 asserted:2 integral:2 xy:1 respective:1 eigenspaces:3 jumping:2 iv:2 walk:2 isolated:9 theoretical:5 column:1 goodness:1 entry:1 kn:6 answer:2 supx:1 st:1 density:4 fundamental:1 siam:1 international:2 stay:2 off:4 together:1 von:6 thesis:1 satisfied:9 worse:1 de:2 matter:1 vi:2 depends:1 view:6 try:2 analyze:1 ulrike:2 start:1 recover:1 complicated:2 parallel:1 ir:6 miller:2 weak:1 iid:1 bisection:1 cybernetics:1 randomness:2 explain:1 definition:3 proof:15 di:9 recovers:1 dataset:1 proved:2 popular:1 treatment:1 recall:1 segmentation:2 actually:1 appears:2 supervised:5 planar:1 wei:4 improved:1 done:2 though:1 shrink:1 hand:1 sketch:1 hopefully:1 lack:1 glance:1 quality:2 behaved:1 usa:1 effect:1 normalized:27 verify:1 true:1 counterpart:1 multiplier:1 hence:7 analytically:1 contain:1 symmetric:2 unnormalized:29 coincides:1 criterion:1 tn:2 performs:1 fj:2 harmonic:1 recently:2 fi:3 common:1 tending:1 empirically:1 overview:1 banach:1 extend:1 assembling:1 interpretation:1 roose:2 ai:2 shuffling:1 meila:3 consistency:8 similarly:2 had:2 dj:1 chapelle:2 similarity:9 add:1 recent:1 showed:1 certain:1 ubingen:1 postponed:1 conserved:1 seen:2 additional:4 care:1 surely:5 converge:16 semi:5 multiple:1 desirable:1 full:1 technical:2 academic:1 laplacian:6 converging:1 basic:2 supi:1 metric:3 vision:1 normalization:2 kernel:1 preserved:1 background:1 want:5 interval:3 completes:1 sch:2 unlike:1 eigenfunctions:3 spemannstr:1 contrary:1 lafferty:1 jordan:1 ideal:3 easy:2 enough:1 xj:13 finish:1 topology:1 simplifies:1 whether:6 linkage:1 wellner:2 pollard:2 useful:1 clear:1 eigenvectors:21 detailed:1 simplest:1 http:1 exist:1 sign:4 disjoint:1 write:1 discrete:1 express:1 group:2 threshold:1 drawn:2 hartigan:2 diffusion:5 graph:5 sum:2 luxburg:7 place:2 almost:5 reasonable:2 extends:1 vn:4 scaling:1 investigates:1 ki:2 guaranteed:1 yale:1 nonnegative:1 annual:1 infinity:2 uii:2 x2:1 bousquet:4 aspect:1 argument:1 vempala:1 relatively:3 px:1 transferred:1 department:2 according:5 ball:1 smaller:1 appealing:1 intuitively:2 multiplicity:13 sij:3 taken:2 ln:22 equation:3 turn:3 discus:1 eventually:1 needed:1 know:2 available:1 away:1 spectral:48 denotes:1 clustering:60 cf:2 ensure:1 guattery:2 unifying:1 mdn:7 unsuitable:1 restrictive:1 ghahramani:1 especially:1 question:5 quantity:1 strategy:3 md:6 diagonal:8 said:1 exhibit:1 dp:4 berlin:1 sensible:2 manifold:1 tuebingen:2 trivial:1 reason:2 kannan:2 length:1 pointwise:4 relationship:1 mini:1 difficult:4 statement:1 debate:1 stated:1 kdn:1 anal:1 unknown:2 rir:1 observation:2 ascoli:1 finite:16 situation:2 extended:1 perturbation:6 arbitrary:2 introduced:1 lal:1 established:1 eigenfunction:4 able:2 usually:2 pattern:1 laplacians:4 reading:1 max:1 reliable:2 including:1 critical:2 difficulty:1 treated:1 natural:1 zhu:2 brief:1 imply:1 lij:5 sn:14 nice:3 multiplication:4 determining:1 interesting:2 suggestion:1 versus:1 foundation:1 degree:6 consistent:2 principle:1 pi:1 balancing:1 row:1 surprisingly:3 uchicago:1 institute:1 fall:1 mikhail:1 van:2 xn:7 dpj:2 valid:1 far:2 cope:1 reconstructed:1 compact:16 preferred:1 dealing:1 global:1 conclude:2 xi:26 spectrum:13 continuous:4 un:20 why:1 additionally:1 inherently:1 separator:1 constructing:2 vj:1 main:3 whole:7 convey:1 ule:1 x1:3 sub:4 sf:3 comput:1 lie:2 vanish:1 third:1 theorem:25 bad:1 load:1 maxi:1 r2:1 dk:1 evidence:2 concern:1 normalizing:1 essential:3 exists:7 workshop:1 phd:1 dissimilarity:1 rg:17 partially:1 driessche:2 collectively:8 springer:2 corresponds:1 satisfies:1 weston:2 vetta:1 identity:2 formulated:1 consequently:1 ann:1 change:3 specifically:1 infinite:1 called:1 teng:2 ksn:1 exception:1 spielman:2 ongoing:1 |
1,864 | 2,693 | Semi-parametric exponential family PCA
Sajama
Alon Orlitsky
Department of Electrical and Computer Engineering
University of California at San Diego, La Jolla, CA 92093
[email protected] and [email protected]
Abstract
We present a semi-parametric latent variable model based technique for
density modelling, dimensionality reduction and visualization. Unlike
previous methods, we estimate the latent distribution non-parametrically
which enables us to model data generated by an underlying low dimensional, multimodal distribution. In addition, we allow the components
of latent variable models to be drawn from the exponential family which
makes the method suitable for special data types, for example binary or
count data. Simulations on real valued, binary and count data show favorable comparison to other related schemes both in terms of separating
different populations and generalization to unseen samples.
1
Introduction
Principal component analysis (PCA) is widely used for dimensionality reduction with applications ranging from pattern recognition and time series prediction to visualization. One
important limitation of PCA is that it is not based on a probability model. A probabilistic formulation of PCA can offer several advantages like allowing statistical testing,
application of Bayesian inference methods and naturally accommodating missing values
[1]. Latent variable models are commonly used in statistics to summarize observations
[2]. A latent variable model assumes that the distribution of data is determined by a latent or mixing
distribution P (?) and a conditional or component distribution P (x|?), i.e.,
R
P (x) = P (?)P (x|?)d?.
Probabilistic PCA (PPCA) [1] borrows from one such popular model called factor analysis to propose a probabilistic alternative PCA. A key feature of this probabilistic model
is that the latent distribution P (?) is also assumed to be Gaussian since it leads to simple
and fast model estimation, i.e., the density of x is approximated by a Gaussian distribution whose covariance matrix is aligned along a lower dimensional subspace. This may
be a good approximation when data is drawn from a single population and the goal is to
explain the data in terms of a few variables. However, in machine learning we often deal
with data drawn from several populations and PCA is used to reduce dimensions to control
computational complexity of learning. A mixture model with Gaussian latent distribution
would not be able to capture this information. The projection obtained using a Gaussian
latent distribution tends to be skewed toward the center [1] and hence the distinction between nearby sub-populations may be lost in the visualization space. For these reasons,
it is important not to make restrictive assumptions about the latent distribution. Several
recently proposed dimension reduction methods can, like PPCA, be thought of as special
cases of latent variable modelling which differ in the speci?c assumptions they make about
the latent and conditional distributions.
We present an alternative probabilistic formulation, called semi-parametric PCA (SPPCA), where no assumptions are made about the distribution of the latent random variable ?. Non-parametric latent distribution estimation allows us to approximate data density
better than previous schemes and hence gives better low dimensional representations. In
particular, multi-modality of the high dimensional density is better preserved in the projected space. When the observed data is composed of several clusters, this technique can
be viewed as performing simultaneous clustering and dimensionality reduction. To make
our method suitable for special data types, we allow the conditional distribution P (x|?) to
be any member of the exponential family of distributions. Use of exponential family distributions for P (x|?) is common in statistics where it is known as latent trait analysis and they
have also been used in several recently proposed dimensionality reduction schemes [3, 4].
We use Lindsay?s non-parametric maximum likelihood estimation theorem to reduce the
estimation problem to one with a large enough discrete prior. It turns out that this choice
gives us a prior which is ?conjugate? to all exponential family distributions, allowing us to
give a uni?ed algorithm for all data types. This choice also makes it possible to ef?ciently
estimate the model even in the case when different components of the data vector are of
different types.
2
The constrained mixture model
We assume that the d-dimensional observation vectors x
R 1 , . . . , xn are outcomes of iid
draws of a random variable whose distribution P (x) = P (?)P (x|?)d? is determined
by the latent distribution P (?) and the conditional distribution P (x|?). This can also be
viewed as a mixture density with P (?) being the mixing distribution, the mixture components labelled by ? and P (x|?) being the component distribution corresponding to ?. The
latent distribution is used to model the interdependencies among the components of x and
the conditional distribution to model ?noise?. For example in the case of a collection of
documents we can think of the ?content? of the document as a latent variable since it cannot
be measured. For any given content, the words used in the document and their frequency
may depend on random factors - for example what the author has been reading recently,
and this can be modelled by P (x|?).
Conditional distribution P (x|?): We assume that P (?) adequately models the dependencies among the components of x and hence that the components of x are independent
when conditioned upon ?, i.e., P (x|?) = ?j P (xj |?j ), where xj and ?j are the j?th components of x and ?. As noted in the introduction, using Gaussian means and constraining
them to a lower dimensional subspace of the data space is equivalent to using Euclidean
distance as a measure of similarity. This Gaussian model may not be appropriate for other
data types, for instance the Bernoulli distribution may be better for binary data and Poisson
for integer data. These three distributions, along with several others, belong to a family
of distributions known as the exponential family [5]. Any member of this family can be
written in the form
log P (x|?) = log P0 (x) + x? ? G(?)
where ? is called the natural parameter and G(?) is a function that ensures that the probabilities sum to one. An important property of this family is that the mean ? of a distribution
and its natural parameter ? are related through a monotone invertible, nonlinear function
? = G0 (?) = g(?). It can be shown that the negative log-likelihoods of exponential family
distributions can be written as Bregman distances (ignoring constants) which are a family
of generalized metrics associated with convex functions [4]. Note that by using different
distributions for the various components of x, we can model mixed data types.
Latent distribution P (?): Like previous latent variable methods, including PCA, we
constrain the latent variable ? to an `-dimensional Euclidean subspace of R d to model the
belief that the intrinsic dimensionality of the data is smaller than d. One way to represent
the (unknown) linear constraint on values that ? can take is to write it as an invertible linear
transformation of another random variable which takes values a ? R` ,
? = aV + b
where V is an ` ? d rotation matrix and b is a d-dimensional displacement vector. Hence
any distribution P? (?) satisfying the low dimensional constraints can be represented using a triple (P (a), V, b), where P (a) is a distribution over R` . Lindsay?s mixture nonparametric maximum likelihood estimation (NPMLE) theorem states that for ?xed (V ,b),
the maximum likelihood (ML) estimate of P (a) exists and is a discrete distribution with no
more than n distinct points of support [6]. Hence if ML is the chosen parameter estimation
technique, the SP-PCA model can be assumed (without loss of generality) to be a constrained ?nite mixture model with at most n mixture components. The number of mixture
components in the model, n, grows with the amount of data and we propose to use pruning
to reduce the number of components during model estimation to help both in computational
speed and model generalization. Finally, we note that instead of the natural parameter, any
of its invertible transformations could have been constrained to a lower dimensional space.
Choosing to linearly constrain the natural parameter affords us computational advantages
similar to those available when we use the canonical link in generalized linear regression.
Low dimensional representation:
There are several ways in which low dimensional
representations can be obtained using the constrained mixture model. We would ideally
like to represent a given observation x by the unknown ? (or the corresponding a related
to ? by ? = aV + b) that generated it, since the conditional distribution P (x|?) is used
to model random effects. However, the actual value of a is not known to us and all of our
knowledge of a is contained in the posterior distribution P (a|x) = P (a)P (x|a)/P (x).
Since a belongs to an `-dimensional space, any of its estimators like the posterior mean
or mode (MAP estimate) can be used to represent x in ` dimensions. For presenting the
simulation results in this paper, we use the posterior mean as the representation point.
This representation has been used in other latent variable methods to get meaningful low
dimensional views [1, 3]. Another method is to represent x by that point ? on (V, b) that
is closest according to the appropriate Bregman distance (it can be shown that there is a
unique such ? opt on the plane). This representation is a generalization of the standard
Euclidean projection and was used in [4].
The Gaussian case: When the exponential family distribution chosen is Gaussian, the
model is a mixture of n spherical Gaussians all of whose means lie on a hyperplane in
the data space. This can be thought of as a ?soft? version of PCA, i.e., Gaussian case of
SP-PCA is related to PCA in the same manner as Gaussian mixture model is related to
K-means. The use of arbitrary mixing distribution over the plane allows us to approximate
arbitrary spread of data along the hyperplane. Use of ?xed variance spherical Gaussians
ensures that like PCA, the direction perpendicular to the plane (V, b) is irrelevant in any
metric involving relative values of likelihoods P (x|? k ), including the posterior mean.
Consider the case when data density P (x) belongs to our model space, i.e., it is speci?ed by
{A, V, b, ?, ?} and let D be any direction parallel to the plane (V, b) along which the latent
distribution P (?) has non-zero variance. Since Gaussian noise with variance ? is added to
this latent distribution to obtain P (x), variance of P (x) along D will be greater than ?.
The variance of P (x) along any direction perpendicular to (V, b) will be exactly ?. Hence,
PCA of P (x) yields the subspace (V, b) which is the same as that obtained using SP-PCA
(this may not be true when P (x) does not belong to our model space). We found that
SP-PCA differs signi?cantly from PPCA in the predictive power of the low-dimensional
density model (see Section 5).
3
Model estimation
Algorithm for ML estimation: We present an EM algorithm for estimating parameters
of a ?nite mixture model with the components constrained to an `-dimensional Euclidean
subspace. We propose an iterative re-weighted least squares (IRLS) method for the maximization step along the lines of generalized linear model estimation. Use of weighted least
squares does not guarantee monotone increase in data likelihood. To guarantee convergence
of the algorithm, we can check the likelihood of data at the IRLS update and decrease step
size if necessary. Let x1 , . . . , xn be iid samples drawn from a d-dimensional density P (x),
c be the number of mixture components and let the mixing density be ? = (?1 , . . . , ?c ).
Associated with each mixture component (indexed by k) are parameter vectors ? k and ak
which are related by ? k = ak V + b. In this section we will work with the assumption
that all components of x correspond to the same exponential family for ease of notation.
For each observed xi there is an unobserved ?missing? variable zi which is a c-dimensional
binary vector whose k?th component is one if the k?th mixture component was the outcome
in the i?th random draw and zero otherwise. If yl is a vector, we use ylm to denote its m?th
component. (Derivation of the algorithm is omitted for lack of space, for details please see
[7]).
The E-step is identical to unconstrained ?nite mixture case,
?k P (xi /? k )
m=1 ?m P (xi /? m )
z?ik = E(zik ) = Pc
;
x
?kj =
In the M-step we update ?, V , b, and ak in the following manner
?k = P n
Pn
?ik
i=1 z
P
c
i=1
m=1 zim
=
Pn
;
[GRi ]l1 =
Pn
?ik xij
i=1 z
P
n
?ik
i=1 z
?ik
i=1 z
n
ai is updated by adding ?ai calculated using
(V ?i V 0 )?ai = GRi
;
[?i ]qq =
?g(?iq )
??iq
d
X
(?
xij ? g(?ij ))Vlj
j=1
Here the function g(?) is as de?ned in Section 2 and depends on the member of the exponential family that is being used. Each column of the matrix V , vs , is updated by adding
?vs calculated using
(A0 ?s A)?vs = GRs
;
[?s ]kk =
?g(?ks )
??ks
;
[GRs ]l1 =
c
X
(?
xk0 s ? g(?k0 s ))Ak0 l
k0 =1
Each component of vector b, bs , is updated by adding ?bs calculated using
Hs ?bs = GRs
;
Hs =
c
X
?g(?k0 s )
??k0 s
0
k =1
;
GRs =
c
X
(?
xk0 s ? g(?k0 s ))
k0 =1
Pruning the mixture components: Redundant mixture components can be pruned between the EM iterations in order to improve speed of the algorithm and generalization
properties while retaining the full capability to approximate P (x). We propose the following criteria for pruning
? Starved components : If ?k < C1 , then drop the k?th component
? Nearby components : If max
i |P (xi |? k1 ) ? P (x| ? k2 )| < C2 , then drop either k1?th
or k2?th component
The value of C1 should be ?(1/n) since we want to measure how starved a component
is based on what percentage of the data it is ?responsible? for. To measure the nearness of
components we use the ?-norm of the difference between probabilities the components
assign to observations since we do not want to lose mixture components that are distinguished with respect to a small number of observation vectors. In the case of clustering
this means that we do not ignore under-represented clusters. C2 should be chosen to be a
small constant, depending on how much pruning is desired.
Convergence of the EM iterations and computational complexity: It is easy to verify
that the SP-PCA model satis?es the continuity assumptions of Theorem 2, [8], and hence
we can conclude that any limit point of the EM iterations is a stationary point of the log
likelihood function. The computational complexity of the E-step is O(cdn) and of the Mstep is O(cd`2 ). For the Gaussian case, the E-step only takes O(c`n) since we only need
to take into account the variation of data along the subspace given by current value of V
(see Section 2). The most expensive step is computation of P (xi |? j ). The k-d tree data
structure is often used to identify relevant mixture components to speed up this step.
Model selection: While any of the standard model selection methods based on penalizing
complexity could be used to choose `, an alternative method is to pick ` which minimizes
a validation or bootstrap based estimate of the prediction error (negative log likelihood per
sample). For the Gaussian case, a fast method to pick ` would be to plot the variance of data
along the principal directions (found using PCA) and look for the dimension at which there
is a ?knee? or a sudden drop in variance or where the total residual variance falls below a
chosen threshold.
Consistency of the Maximum Likelihood estimator: We propose to use the ML estimator to ?nd the latent space (V, b) and the latent distribution P (a). Usually a parametric
form is assumed for P (a) and the consistency of the ML estimate is well known for this
task where the parameter space is a subset of a ?nite dimensional Euclidean space. In
the SP-PCA model, one of the parameters (P (a)) ranges over the space of all distribution
functions on R` and hence we need to do more to verify the validity of our estimator. Exponential family mixtures are not identi?able in general. This, however, is not a problem
for us since we are only interested in approximating P (x) well and not in the actual parameters corresponding to the distribution. Hence we use the de?nition of consistency of
an estimator given by Redner. Let ?0 be the ?true? parameter from which observed samples
are drawn. Let C0 be the set of all parameters ? corresponding to the ?true? distribution
F (x/?0 ) (i.e., C0 = {? : F (x/?) = F (x/?0 ) ? x}). Let ??n be an estimator of ? based
? be the quotient topological space obtained from ?
on n observed samples of X and let ?
obtained by identifying the set C0 to a point ??0 .
De?nition The sequence of estimators {?
? n , n = 1, . . . , ?} is said to be strongly
consistent in the sense of Redner if limm?? ??n = ??0 almost surely.
Theorem If P (a) is assumed to be zero outside a bounded subset of R` , the ML estimator of parameter (V, b, P (a)) is strongly consistent for Gaussian, Binary and Poisson
conditional distributions.
The theorem follows by verifying that the assumptions of Kiefer et. al. [9] are satis?ed
by the SP-PCA model. The assumption that P (a) is zero outside a bounded region is not
restrictive in practice since we expect the observations xi belong to a bounded region of
Rd . (Proof omitted for lack of space, please see [7]).
Table 1: Bootstrap estimates of prediction error for PPCA and SP-PCA.
D ENSITY
ERROR
4
I SOTROPIC
GAUSSIAN
`=1
PPCA
`=2
`=3
`=1
SP-PCA
`=2
`=3
F ULL
GAUSSIAN
50.39
38.03
34.71
34.76
36.85
30.99
28.54
343.83
Relationship to past work
SP-PCA is a factor model that makes fewer assumptions about latent distribution than
PPCA [1]. Mixtures of probabilistic principal component analyzers (also known as mixtures of factor analyzers) is a generalization of PPCA which overcomes the limitation of
global linearity of PCA via local dimensionality reduction. Mixtures of SP-PCA?s can be
similarly de?ned and used for local dimensionality reduction. Collins et. al. [4] proposed a
generalization of PCA using exponential family distributions. Note that this generalization
is not associated with a probability density model for the data. SP-PCA can be thought of
as a ?soft? version of this generalization of PCA, in the same manner as Gaussian mixtures
are a soft version of K-means. Generative topographic mapping (GTM) is a probabilistic
alternative to Self organizing map which aims at ?nding a nonlinear lower dimensional
manifold passing close to data points. An extension of GTM using exponential family distributions to deal with binary and count data is described in [3]. Apart from the fact that
GTM is a non-linear dimensionality reduction technique while SP-PCA is globally linear
like PCA, one main feature that distinguishes the two is the choice of latent distribution.
GTM assumes that the latent distribution is uniform over a ?nite and discrete grid of points.
Both the location of the grid and the nonlinear mapping are to be given as an input to the
algorithm. Tibshirani [10] used a semi-parametric latent variable model for estimation of
principle curves. Discussion of these and other dimensionality reduction schemes based on
latent trait and latent class models can be found in [7].
5
Experiments
In this section we present simulations on synthetic and real data to demonstrate the properties of SP-PCA. In factor analysis literature, it is commonly believed that choice of prior
distribution is unimportant for the low dimensional data summarization (see [2], Sections
2.3, 2.10 and 2.16). Through the examples below we argue that estimating the prior instead
of assuming it arbitrarily can make a difference when latent variable models are used for
density approximation, data analysis and visualization.
Use of SP-PCA as a low dimensional density model:
The Tobamovirus data which
consists of 38 18-dimensional examples was used in [1] to illustrate properties of PPCA.
PPCA and SP-PCA can be thought of as providing a range of low-dimensional density
models for the data. The complexity of these densities increases with and is controlled by
the value of ` (the projected space dimension) starting with the zero dimensional model of
an isotropic Gaussian. For a ?xed lower dimension `, SP-PCA has greater approximation
capability than PPCA. In Table 1, we present bootstrap estimates of the predictive power of
PPCA and SP-PCA for various values of L. SP-PCA has lower prediction error than PPCA
for ` = 1, 2 and 3. This indicates that SP-PCA combines ?exible density estimation and
excellent generalization even when trained on a small amount of data.
Simulation results on discrete datasets:
We present experiments on 20 Newsgroups
dataset comparing SP-PCA to PCA, exponential family GTM [3] and Exponential family
PCA [4]. Data for the ?rst set of simulations was drawn from comp.sys.ibm.pc.hardware,
comp.sys.mac.hardware and sci.med newsgroups. A dictionary size of 150 words was
chosen and the words in the dictionary were picked to be those which have maximum
mutual information with class labels. 200 documents were drawn from each of the three
newsgroups to form the training data. Two-dimensional representations obtained using
various methods are shown in Fig. 1. In the projection obtained using PCA, Exponential
family PCA and Bernoulli GTM, the classes comp.sys.ibm.pc.hardware and comp.sys.mac.hardware were not well separated in the 2D space. This result (Fig. 1(c)) was presented
in [3] and the the overlap between the two groups was attributed to the fact that they are
very similar and hence share many words in common. However, SP-PCA was able to
separate the three sets reasonably well (Fig. 1(d)). One way to quantify the separation of
dissimilar groups in the two-dimensional projections is to use the training set classi?cation
error of projected data using SVM. The accuracy of the best SVM classi?er (we tried a
range of SVM parameter values and picked the best for each projected data set) was 75% for
bernoulli GTM projection and 82.3% for SP-PCA projection (the difference corresponds to
44 data points while the total number of data points is 600). We conjecture that the reason
comp.sys.ibm.pc.hardware and comp.sys.mac.hardware have overlap in projection using
Bernoulli GTM is that the prior is assumed to be over a pre-speci?ed grid in latent space
and the spacing between grid points happened to be large in the parameter space close to the
two news groups. In contrast to this, in SP-PCA there is no grid and the latent distribution
is allowed to adapt to the given data set. Note that a standard clustering algorithm could be
used on the data projected using SP-PCA to conclude that data consisted of three kinds of
documents.
20
1
1
100
0.8
10
0.5
0.6
0
80
0.4
0
60
0.2
?0.5
0
?10
?1
40
?0.2
?20
20
?0.4
?1.5
?0.6
0
?2
?30
?0.8
?2.5
?20
?1
?1.5
?1
?0.5
0
0.5
1
1.5
(a) PCA
2
2.5
3
?40
?50
?40
?30
?20
?10
0
10
(b) Expontl. PCA
20
?1
?0.8
?0.6
?0.4
?0.2
0
0.2
0.4
(c) GTM
0.6
0.8
1
?60
?40
?20
0
20
40
60
80
100
(d) SP-PCA
Figure 1: Projection by various methods of binary data from 200 documents each from
comp.sys.ibm.pc.hardware (?), comp.sys.mac.hardware (?) and sci.med (.)
Data for the second set of simulations was drawn from sci.crypt, sci.med, sci.space and
soc.culture.religion.christianity newsgroups. A dictionary size of 100 words was chosen
and again the words in the dictionary were picked to be those which have maximum mutual information with class labels. 100 documents were drawn from each of the newsgroups
to form the training data and 100 more to form the test data. Fig. 2 shows two-dimensional
representations of binary data obtained using various methods. Note that while the four
newsgroups are bunched together in the projection obtained using Exponential family PCA
[4] (Fig. 2(b)), we can still detect the presence four groups from this projection and in
this sense this projection is better than the PCA projection. This result is pleasing since it
con?rms our intuition that using negative log-likelihood of Bernoulli distribution as a measure of similarity is more appropriate than squared Euclidean distance for binary data. We
conjecture that the reason the four groups are not well separated in this projection is that
a conjugate prior has to be used in its estimation for computational purposes [4] and the
form and parameters of this prior are considered ?xed and given inputs to the algorithm.
Both SP-PCA (Fig. 2(c)) and Bernoulli GTM (Fig. 2(e)) were able to clearly separate the
clusters in the training data. Figures 2(d) and 2(f) show representation of test data using the
models estimated by SP-PCA and Bernoulli GTM respectively. To measure generalization
of these methods, we use a K-nearest neighbors based non-parametric estimate of the density of the projected training data. The percentage difference between the log-likelihoods
of training and test data with respect to this density was 9.1% for SP-PCA and 17.6% for
GTM for K=40 (SP-PCA had smaller percentage change in log-likelihood for most values
of K that we tried between 10 and 40). This indicates that SP-PCA generalizes better than
GTM. This can be seen visually by comparing Figures 2(e) and 2(f) where the projections
of training and test data of sci.space (?) differ signi?cantly.
1
14
0.5
12
0
50
10
0
?0.5
8
?1
?50
6
?1.5
4
?100
?2
2
?2.5
?150
?3
?4
?3
?2
?1
0
1
2
0
?20
?15
(a) PCA
?10
?5
0
5
10
15
20
25
30
?200
?150
(b) Exponential PCA
50
?100
?50
0
(c) SP-PCA
1
1
0.8
0.8
0.6
0.6
0.4
0.4
0
0.2
0.2
0
0
?0.2
?0.2
?0.4
?0.4
?0.6
?0.6
?50
?100
?0.8
?0.8
?150
?1
?200
?150
?100
?50
(d) Test data - SP-PCA
0
?1
?1
?0.8
?0.6
?0.4
?0.2
0
0.2
0.4
0.6
(e) Bernoulli GTM
0.8
1
?1
?0.8
?0.6
?0.4
?0.2
0
0.2
0.4
0.6
0.8
1
(f) Test data - GTM
Figure 2: Projection by various methods of binary data from 100 documents each from
sci.crypt (?), sci.med (?), sci.space (?) and soc.culture.religion.christianity (+)
Acknowledgments
We thank Sanjoy Dasgupta and Thomas John for helpful conversations.
References
[1] M. Tipping and C. Bishop. Probabilistic principal component analysis. Journal of the Royal
Statistical Society, Series B, 61(3):611?622, 1999.
[2] David J. Bartholomew and Martin Knott. Latent variable models and Factor analysis, volume 7
of Kendall?s Library of Statistics. Oxford University Press, 2nd edition, 1999.
[3] A. Kaban and M. Girolami. A combined latent class and trait model for the analysis and visualization of discrete data. IEEE Transaction on Pattern Analysis and Machine Intelligence,
23(8):859?872, August 2001.
[4] M. Collins, S. Dasgupta, and R. E. Schapire. A generalization of principal components analysis
to the exponential family. In Advances in Neural Information Processing Systems 14, 2002.
[5] P. McCullagh and J. A. Nelder. Generalized Linear Models. Monographs on Statistics and
Applied Probability. Chapman and Hall, 1983.
[6] B. G. Lindsay. The geometry of mixture likelihoods : A general theory. The Annals of Statistics,
11(1):86?04, 1983.
[7] Sajama and A. Orlitsky. Semi-parametric exponential family PCA : Reducing dimensions via
non-parametric latent distribution estimation. Technical Report CS2004-0790, University of
California at San Diego, http://cwc.ucsd.edu/? sajama, 2004.
[8] C. F. J. Wu. On the convergence properties of the EM algorithm. Annals of Statistics, 11(1):95?
103, 1983.
[9] J. Kiefer and J. Wolfowitz. Consistency of the maximum likelihood estimator in the presence
of in?nitely many incidental parameters. The Annals of Mathematical Statistics, 27:887?906,
1956.
[10] R. Tibshirani. Principal curves revisited. Statistics and Computation, 2:183?190, 1992.
| 2693 |@word h:2 version:3 norm:1 nd:2 c0:3 simulation:6 tried:2 covariance:1 p0:1 pick:2 reduction:9 series:2 document:8 past:1 current:1 comparing:2 written:2 john:1 enables:1 mstep:1 drop:3 plot:1 update:2 zik:1 v:3 stationary:1 fewer:1 generative:1 intelligence:1 plane:4 isotropic:1 sys:8 sudden:1 nearness:1 revisited:1 location:1 mathematical:1 along:9 c2:2 ik:5 consists:1 combine:1 manner:3 multi:1 globally:1 spherical:2 actual:2 estimating:2 underlying:1 notation:1 bounded:3 linearity:1 what:2 xed:4 kind:1 minimizes:1 unobserved:1 transformation:2 guarantee:2 orlitsky:2 exactly:1 ull:1 k2:2 control:1 engineering:1 local:2 tends:1 limit:1 ak:3 oxford:1 nitely:1 k:2 ease:1 perpendicular:2 range:3 unique:1 responsible:1 acknowledgment:1 testing:1 lost:1 practice:1 differs:1 bootstrap:3 nite:5 displacement:1 thought:4 projection:15 word:6 pre:1 get:1 cannot:1 close:2 selection:2 equivalent:1 map:2 missing:2 center:1 starting:1 convex:1 christianity:2 knee:1 identifying:1 estimator:9 population:4 variation:1 updated:3 qq:1 diego:2 lindsay:3 annals:3 recognition:1 approximated:1 satisfying:1 expensive:1 observed:4 electrical:1 capture:1 verifying:1 region:2 ensures:2 news:1 decrease:1 monograph:1 intuition:1 complexity:5 ideally:1 trained:1 depend:1 predictive:2 upon:1 multimodal:1 k0:6 various:6 represented:2 gtm:15 derivation:1 separated:2 distinct:1 fast:2 outcome:2 choosing:1 outside:2 whose:4 widely:1 valued:1 otherwise:1 statistic:8 unseen:1 topographic:1 think:1 advantage:2 sequence:1 propose:5 aligned:1 relevant:1 organizing:1 mixing:4 rst:1 convergence:3 cluster:3 help:1 iq:2 alon:2 depending:1 illustrate:1 measured:1 nearest:1 ij:1 soc:2 quotient:1 signi:2 quantify:1 differ:2 direction:4 girolami:1 bunched:1 assign:1 generalization:11 opt:1 extension:1 considered:1 hall:1 visually:1 mapping:2 dictionary:4 omitted:2 purpose:1 favorable:1 estimation:14 lose:1 label:2 ylm:1 weighted:2 clearly:1 gaussian:18 aim:1 pn:3 zim:1 modelling:2 likelihood:15 bernoulli:8 check:1 indicates:2 contrast:1 sense:2 detect:1 helpful:1 inference:1 a0:1 limm:1 interested:1 among:2 retaining:1 constrained:5 special:3 mutual:2 chapman:1 identical:1 look:1 others:1 report:1 few:1 distinguishes:1 composed:1 geometry:1 pleasing:1 satis:2 mixture:26 pc:5 bregman:2 necessary:1 culture:2 indexed:1 tree:1 euclidean:6 re:1 desired:1 instance:1 column:1 soft:3 maximization:1 mac:4 parametrically:1 subset:2 uniform:1 sajama:4 dependency:1 synthetic:1 combined:1 density:17 cantly:2 probabilistic:8 yl:1 invertible:3 together:1 again:1 squared:1 choose:1 account:1 de:4 depends:1 view:1 picked:3 kendall:1 parallel:1 capability:2 square:2 kiefer:2 accuracy:1 variance:8 yield:1 correspond:1 identify:1 modelled:1 bayesian:1 irls:2 iid:2 comp:8 cation:1 explain:1 simultaneous:1 ed:4 crypt:2 frequency:1 naturally:1 associated:3 proof:1 attributed:1 con:1 ppca:12 dataset:1 popular:1 knowledge:1 conversation:1 dimensionality:9 redner:2 tipping:1 formulation:2 strongly:2 generality:1 nonlinear:3 lack:2 continuity:1 mode:1 grows:1 effect:1 validity:1 verify:2 true:3 consisted:1 adequately:1 hence:10 deal:2 skewed:1 during:1 self:1 please:2 noted:1 criterion:1 generalized:4 presenting:1 demonstrate:1 l1:2 ranging:1 ef:1 recently:3 common:2 rotation:1 volume:1 belong:3 trait:3 ai:3 rd:1 unconstrained:1 consistency:4 grid:5 similarly:1 analyzer:2 bartholomew:1 had:1 similarity:2 posterior:4 closest:1 jolla:1 belongs:2 irrelevant:1 apart:1 binary:10 arbitrarily:1 nition:2 seen:1 greater:2 speci:3 surely:1 wolfowitz:1 redundant:1 semi:5 full:1 interdependency:1 technical:1 adapt:1 offer:1 believed:1 ensity:1 controlled:1 prediction:4 involving:1 regression:1 metric:2 poisson:2 iteration:3 represent:4 c1:2 preserved:1 addition:1 want:2 spacing:1 modality:1 unlike:1 med:4 member:3 integer:1 ciently:1 presence:2 constraining:1 enough:1 easy:1 newsgroups:6 xj:2 zi:1 gri:2 reduce:3 pca:62 rms:1 passing:1 unimportant:1 amount:2 nonparametric:1 hardware:8 schapire:1 http:1 affords:1 xij:2 canonical:1 percentage:3 happened:1 estimated:1 per:1 tibshirani:2 discrete:5 write:1 dasgupta:2 group:5 key:1 four:3 threshold:1 drawn:9 penalizing:1 monotone:2 sum:1 family:23 almost:1 wu:1 separation:1 draw:2 topological:1 constraint:2 constrain:2 nearby:2 speed:3 pruned:1 performing:1 martin:1 conjecture:2 ned:2 department:1 according:1 conjugate:2 smaller:2 em:5 b:3 visualization:5 turn:1 count:3 available:1 gaussians:2 generalizes:1 appropriate:3 distinguished:1 alternative:4 thomas:1 assumes:2 clustering:3 restrictive:2 k1:2 approximating:1 society:1 g0:1 added:1 parametric:10 said:1 subspace:6 distance:4 link:1 separate:2 separating:1 sci:9 thank:1 accommodating:1 manifold:1 argue:1 toward:1 reason:3 assuming:1 relationship:1 kk:1 providing:1 negative:3 incidental:1 xk0:2 summarization:1 unknown:2 allowing:2 av:2 observation:6 datasets:1 knott:1 ucsd:3 arbitrary:2 august:1 david:1 california:2 identi:1 distinction:1 able:4 below:2 pattern:2 usually:1 reading:1 summarize:1 including:2 max:1 royal:1 belief:1 power:2 suitable:2 overlap:2 natural:4 residual:1 scheme:4 improve:1 library:1 nding:1 kj:1 prior:7 literature:1 relative:1 loss:1 expect:1 mixed:1 limitation:2 borrows:1 triple:1 cdn:1 validation:1 consistent:2 principle:1 share:1 cd:1 ibm:4 allow:2 fall:1 neighbor:1 curve:2 dimension:7 xn:2 calculated:3 ak0:1 sppca:1 author:1 commonly:2 made:1 san:2 projected:6 collection:1 transaction:1 approximate:3 pruning:4 uni:1 ignore:1 kaban:1 overcomes:1 ml:6 global:1 assumed:5 conclude:2 nelder:1 xi:6 latent:38 iterative:1 table:2 reasonably:1 ca:1 ignoring:1 excellent:1 sp:33 spread:1 main:1 linearly:1 noise:2 edition:1 allowed:1 x1:1 fig:7 sub:1 exponential:20 lie:1 theorem:5 bishop:1 exible:1 er:1 svm:3 starved:2 intrinsic:1 exists:1 adding:3 conditioned:1 contained:1 religion:2 corresponds:1 conditional:8 goal:1 viewed:2 labelled:1 content:2 change:1 mccullagh:1 determined:2 reducing:1 hyperplane:2 vlj:1 classi:2 principal:6 called:3 total:2 sanjoy:1 ece:1 e:1 la:1 meaningful:1 support:1 collins:2 dissimilar:1 |
1,865 | 2,694 | Saliency-Driven Image Acuity Modulation on a
Reconfigurable Silicon Array of Spiking Neurons
R. Jacob Vogelstein1 , Udayan Mallik2 , Eugenio Culurciello3 ,
Gert Cauwenberghs2 and Ralph Etienne-Cummings2
1
Dept. of Biomedical Engineering, Johns Hopkins University, Baltimore, MD
2
Dept. of Electrical & Computer Engineering, Johns Hopkins University, Baltimore, MD
3
Dept. of Electrical Engineering, Yale University, New Haven, CT
{jvogelst,umallik1,gert,retienne}@jhu.edu, [email protected]
Abstract
We have constructed a system that uses an array of 9,600 spiking silicon neurons, a fast microcontroller, and digital memory, to implement
a reconfigurable network of integrate-and-fire neurons. The system is
designed for rapid prototyping of spiking neural networks that require
high-throughput communication with external address-event hardware.
Arbitrary network topologies can be implemented by selectively routing address-events to specific internal or external targets according to a
memory-based projective field mapping. The utility and versatility of the
system is demonstrated by configuring it as a three-stage network that
accepts input from an address-event imager, detects salient regions of the
image, and performs spatial acuity modulation around a high-resolution
fovea that is centered on the location of highest salience.
1
Introduction
The goal of neuromorphic engineering is to design large-scale sensory information processing systems that emulate the brain. In many biological neural systems, the information
received by a sensory organ passes through multiple stages of neural computations before
a judgment is made. A convenient way to study this functionality is to design separate
chips for each stage of processing and connect them with a fast data bus. However, it is not
always advisable to fabricate a new chip to test a hypothesis regarding a particular neural
computation, and software models of spiking neural networks cannot typically execute or
communicate with external devices in real-time. Therefore, we have designed specialized
hardware that implements a reconfigurable array of spiking neurons for rapid prototyping
of large-scale neural networks.
Neuromorphic sensors can generate up to millions of spikes per second (see, e.g., [1]),
so a proper communication protocol is required for multi-chip systems. ?Address-Event
Representation? (AER) was developed for this purpose over a decade ago and has since
become the common ?language? of neuromorphic chips [2?7]. The central idea of AER
is to use time-multiplexing to emulate extensive connectivity between neurons. Although
it was originally proposed to implement a one-to-one connection topology, AER has been
extended to allow convergent and divergent connectivity [5, 8, 9], and has even been used
DAC
driving potential
(analog)
DIO
MCU
driving
potential
(digital)
postsynaptic
neuron addr.
presynaptic
neuron addr.
RAM
synapse
parameters
postsynaptic
neuron addr.
MCU
I&F
synaptic
weight
outgoing
spike addr.
incoming
spike addr.
I&F
recurrent
mode
DIO
(a)
feed-forward
mode
RAM
DAC
(b)
Figure 1: (a) Block diagram of IFAT system. Incoming and outgoing address-events are
communicated through the digital I/O port (DIO), with handshaking executed by the microcontroller (MCU). The digital-to-analog converter (DAC) is controlled by the MCU and
provides the synaptic driving potential (?E? in Figure 2) to the integrate-and-fire neurons
(I&F), according to the synapse parameters stored in memory (RAM). Modified from [18].
(b) Printed circuit board integrating all components of the IFAT system.
for functions in addition to inter-chip communication [10?12]. Within our hardware array, all inter-neuron communication is performed using AER; the absence of hardwired
connections is the feature that allows for reconfigurability.
A few examples of AER-based reconfigurable neural array transceivers can be found in the
literature [8, 9], but our Integrate-and-Fire Array Transceiver system (IFAT) differs in its
size and flexibility. With four custom aVLSI chips [13] operating in parallel and 128 MB
of digital RAM, the system contains 9,600 neurons and up to 4,194,304 synapses. Because
it was designed from the start for generality and biological realism, every silicon neuron
implements a discrete-time version of the classical biological ?membrane equation? [14],
a simple conductance-like model of neural function that allows for emulating an unlimited
number of synapse types by dynamically varying two parameters [13]. By using a memorybased projective field mapping to route incoming address-events to different target neurons,
the system can implement arbitrarily complex network topologies, limited only by the capacity of the RAM.
To demonstrate the functionality of the IFAT, we designed a three-stage feed-forward model
of salience-based attention and implemented it entirely on the reconfigurable array. The
model is based on a biologically-plausible architecture that has been used to explain human
visual search strategies [15, 16]. Unlike previous hardware implementations (e.g. [17]), we
use a multi-chip system and perform all computations with spiking neurons. The network
accepts spikes from an address-event imager as inputs, computes spatial derivatives of light
intensity as a measure of local information content, identifies regions of high salience, and
foveates a location of interest by reducing the resolution in the surrounding areas. These
capabilities are useful for smart, low-bandwidth, wide-angle surveillance networks.
2
Hardware
From the perspective of an external device, the IFAT system (Figure 1) operates as an
AER transceiver, both receiving and transmitting spikes over a bidirectional address-event
(AE) bus. Internally, incoming events are routed to any number of integrate-and-fire (I&F)
Vdd
?2
?1
Vbp
M5
Vdd
Vcomp
Rscan
M11
Vm
E
X1
Creq
Vdd
M6
M15
M7 M8
M4
C1
C2
M12
?1
?2
W0
M1
W1
M2
W2
M3
Rreq
M14
X2
C0
M13
Cack
Cm
M9
Rack
M10
Vthresh
Vbn
Vreset
Figure 2: Silicon neuron. The ?general-purpose? synapse is shown inside the dashed
box [13], with event generation circuitry shown right [9].
neurons according to a look-up table stored in RAM. When the inputs are sufficient to
cause a neuron to spike, the output is either directed to other internal neurons (for recurrent
networks) or to an external device via the AE bus. The following two sections will describe
the system and silicon neurons in more detail.
2.1
The IFAT system
A block diagram of the IFAT system and its physical implementation are shown in Figure 1.
The primary components are a 100 MHz FPGA (Xilinx XC2S100PQ208), 128 MB of nonvolatile SRAM (TI bq4017), a high-speed DAC (TI TLC7524CN), a 68-pin digital I/O
interface (DIO), and 4 custom aVLSI chips that implement a total of 9,600 I&F neurons.
The FPGA controls access to both an internal and external AE bus, and communicates
address-events between both the I&F neurons and external devices in bit-parallel using a
four-phase asynchronous handshaking scheme.
The 128 MB of RAM is arranged as a 4 MB ? 32 array. Each 32-bit entry contains a
complete description of a single synapse, specifying the postsynaptic target, the synaptic
equilibrium potential, and the synaptic weight. The weight field can be further subdivided
into three parts, corresponding to three ways in which biological neurons can control the
synaptic weight (w) [14, p. 91]:
w = npq
(1)
where n is the number of quantal neurotransmitter sites, p is the probability of neurotransmitter release per site, and q is a measure of the postsynaptic effect of the neurotransmitter.
In the IFAT system, the FPGA can implement p with a simple pseudo-random number algorithm, it can control n by sending multiple outgoing spikes for each incoming spike, and
it sends the value of q to the I&F neuron chips (see Section 2.2).
Instead of hardwired connections between neurons, the IFAT implements ?virtual connections? by serially routing incoming events to their appropriate targets at a rate of up to
1,000,000 events per second. When the IFAT receives an AE from an external device, the
FPGA observes the address, appends some ?chip identifier? (CID) bits, and stores the resulting binary number as a base address. It then adds additional offset bits to form
a complete 22-bit RAM address, which it uses to look up a set of synaptic parameters.
After configuring q and instructing the DAC to produce the analog synaptic equilibrium
potential, the FPGA activates a target neuron by placing its address on the internal AE
bus and initiating asynchronous handshaking with the appropriate I&F chip. It then increments the offset by one and repeats the process for the next synapse, stopping when it
sees a reserved code word in the data field. Recurrent connections can be implemented
simply by appending a different CID to events generated by the on-board I&F neurons,
while connections to external devices are achieved by specifying the appropriate CID for
Row Decoder
E (V) W
Column Decoder
Spike Threshold
V
m
1.5
1
60 x 40 Array of Neurons
0.5
Column Arbitration
111
000
5
0
0
0.1
0.2
0.3
Time (Arbitrary Units)
0.4
(a)
Row Arbitration
LFSR
0
(b)
Figure 3: (a) Data collected from one neuron during operation of the chip. The lower trace
illustrates the membrane potential (Vm ) of a single neuron in the array as a series of events
are sent at times marked at the bottom of the figure. The synaptic equilibrium potential (E)
and synaptic weight (W ) are drawn in the top two traces. Figure from [13]. (b) Integrateand-fire chip micrograph. The linear-feedback shift register (LFSR) implements a pseudorandom element for resolving arbitration conflicts. Modified from [13].
the postsynaptic target. With this infrastructure, arbitrary patterns of connectivity can be
implemented, limited only by the memory?s capacity.
2.2
Integrate-and-Fire Neurons
As described above, the IFAT system includes four custom aVLSI chips [13] that contain a
total of 9,600 integrate-and-fire neurons. All the neurons are identical and each implements
a simple conductance-like model of a single, ?general purpose? synapse using a switchedcapacitor architecture (Figure 2). The synapses have two internal parameters that can be
dynamically modulated for each incoming event: the synaptic equilibrium potential (E)
and the synaptic weight (W0-W2). Values for both parameters are stored in RAM; the
3-bit q is used by the FPGA to selectively enable binary-sized capacitors C0-C2, while E
is converted to an analog value by the DAC. By varying these parameters, it is possible
to emulate a large number of different kinds of synapses impinging on the same cell. An
example of one neuron in operation is shown in Figure 3a.
A micrograph of the integrate-and-fire chip is shown in Figure 3b. Incoming address-events
are decoded and sent to the appropriate neuron in the 60 ? 40 array. When a neuron?s
membrane potential exceeds an externally-provided threshold voltage, it requests service
from the peripheral arbitration circuitry. After request is acknowledged, the neuron is reset and its address is placed on the IFAT system?s internal AER bus. Conflicts between
simultaneously active neurons are resolved by a novel arbitration scheme that includes a
pseudo-random element on-chip [19].
3
Experimental Design and Results
To demonstrate the functionality of the IFAT system, we designed and implemented a threestage network for salience-based foveation [16] of an address-event imager. This work
is motivated by the fact that wide-angle image sensors in a monitoring sensor network
(a)
(b)
Figure 4: (a) Test image. (b) Output from Octopus Retina.
extract a large quantity of data from the environment, most of which is irrelevant. Because
bandwidth is limited and data transmission is energy-intensive, it is desirable to reduce the
amount of information sent over the communication channel. Therefore, if a particular
region of the visual field can be identified as having high salience, that part of the image
can be selectively transmitted with high resolution and the surrounding scene can be be
compressed.
The input to the first stage of the network is a stream of address-events generated by an
asynchronous imager called the ?Octopus Retina? (OR) [1,20]. The OR contains a 60 ? 80
array of light-sensitive ?neurons? that each represent local light intensity as a spike rate. In
other words, pixels that receive a lot of light spike more frequently than those that receive
a little light. For these experiments, we collected 100,000 events from the OR over the
course of about one second while it was viewing a grayscale picture mounted on a white
background. The test image and OR output are shown in Figure 4.
To identify candidate regions of salience, the first stage of the network is configured to
compute local changes in contrast. Every 2 ? 2 block of pixels in the OR corresponds
to four neurons on the IFAT that respond to light-to-dark or dark-to-light transitions in
the rightward or downward direction (Figure 5a). Each IFAT cell computes local changes
in contrast due to a receptive field (RF) that spans four OR pixels in either the horizontal or vertical dimension, with two of its inputs being excitatory and the other two being
inhibitory. When a given IFAT cell?s RF falls on a region of visual space with uniform
brightness, all of the OR pixels projecting to that cell will have the same mean firing rate,
so the excitatory and inhibitory inputs will cancel. However, if a cell?s excitatory inputs
are exposed to high light intensity and its inhibitory inputs are exposed to low light intensity, the cell will receive more excitatory inputs than inhibitory inputs and will generate an
output spike train with spike frequency proportional to the contrast. The output from the
4,800 IFAT neurons in the first stage of the network in response to the OR input is shown
in Figure 5b.
The second stage of processing is designed to pool inputs from neighboring contrastsensitive cells to identify locations of high salience. Our underlying assumption is that
regions of interest will contain more detail than their surroundings, producing a large output from the first stage. Blocks of 8 ? 8 IFAT cells from the first stage project to single
cells in the second stage, and each 8 ? 8 region overlaps the next by 4 neurons (Figure 6a).
Therefore, every IFAT cell in the second stage has an 8 ? 8 RF. Although it is not necessary to normalize the firing rates of the first and second stages, because every second stage
IFAT cell receives 64 inputs, we reduce the strength of the synaptic connections between
the two stages to conserve bandwidth. The output from the 300 IFAT neurons in the second
1
4
2
3
(a)
(b)
Figure 5: (a) Stage 1 network for computing local changes in contrast. Squares in the
center represent OR pixels. Circles represent IFAT neurons. Excitatory synapses are represented by triangles, and inhibitory synapses as circles. Only four IFAT neurons with nonoverlapping receptive fields are shown for clarity. (b) Output of stage 1, as implemented on
the IFAT, with Figure 4b from the OR as input.
stage of the network in response to the output from the first stage IFAT neurons is shown
in Figure 6b.
The final stage of processing modulates the spatial acuity of the original image to reduce the
resolution outside the region of highest salience. This is achieved by a foveation network
that pools inputs from neighboring pixels using overlapping Gaussian kernels (Figure 7a)
[18]. The shape of the kernel functions is implemented by varying the synaptic weight
and synaptic equilibrium potential between OR neurons and IFAT cells in the third stage:
within every pooled block, the strongest connections originate from the center pixels and
the weakest connections come from the outermost pixels. Instead of physically moving the
OR imager to center the fovea on the region of interest, we relocate the fovea by performing
simple manipulations in the address domain. First, the address space of incoming events
is enlarged beyond the range provided by the OR and the fovea is centered within this
virtual visual field (Figure 7a). Then, the row and column address of the second stage IFAT
neuron with the largest output is subtracted from the address of the center of the fovea,
and the result is stored as a constant offset. This offset is then added to the addresses of
all incoming events from the OR, resulting in a shift of the OR image in the virtual visual
field so that the fovea will be positioned over the region of highest salience. The output
from the 1,650 IFAT neurons in the third stage network is shown in Figure 7b. With a
32 ? 32 pixel high-resolution fovea, the network allows for a 66% reduction in the number
of address-events required to reconstruct the image.
4
Conclusion
We have demonstrated a multi-chip neuromorphic system for performing saliency-based
spatial acuity modulation. An asynchronous imager provides the input and communicates
with a reconfigurable array of spiking silicon neurons using address-events. The resulting
output is useful for efficient spatial and temporal bandwidth allocation in low-power vision
sensors for wide-angle video surveillance. Future work will concentrate on extending the
functionality of the multi-chip system to perform stereo processing on address-event data
from two imagers.
(a)
(b)
Figure 6: (a) Stage 2 network for computing local changes in contrast. Blocks of 8 ? 8
IFAT neurons from stage 1 (shown as regions alternately shaded white and gray) project
to single IFAT neurons in stage 2 (not shown). Blocks are shown as non-overlapping for
clarity. (b) Output of stage 2, as implemented on the IFAT, with Figure 5b from stage 1 as
input.
Acknowledgments
This work was partially funded by NSF Awards #0120369, #9896362, and IIS-0209289;
ONR Award #N00014-99-1-0612; and a DARPA/ONR MURI #N00014-95-1-0409. Additionally, RJV is supported by an NSF Graduate Research Fellowship.
References
[1] E. Culurciello, R. Etienne-Cummings, and K. A. Boahen, ?A biomorphic digital image sensor,?
IEEE J. Solid-State Circuits, vol. 38, no. 2, 2003.
[2] M. Sivilotti, Wiring considerations in analog VLSI systems, with application to fieldprogrammable networks. PhD thesis, California Institute of Technology, Pasadena, CA, 1991.
[3] M. Mahowald, An analog VLSI system for stereoscopic vision. Boston, MA: Kluwer Academic
Publishers, 1994.
[4] J. Lazzaro, J. Wawrzynek, M. Mahowald, M. Sivilotti, and D. Gillespie, ?Silicon auditory processors as computer peripherals,? IEEE Trans. Neural Networks, vol. 4, no. 3, pp. 523?528,
1993.
[5] K. A. Boahen, ?Point-to-point connectivity between neuromorphic chips using address events,?
IEEE Trans. Circuits & Systems II, vol. 47, no. 5, pp. 416?434, 2000.
[6] C. M. Higgins and C. Koch, ?Multi-chip neuromorphic motion processing,? in Proc. 20th Anniversary Conference on Advanced Research in VLSI (D. S. Wills and S. P. DeWeerth, eds.),
(Los Alamitos, CA), pp. 309?323, IEEE Computer Society, 1999.
[7] S.-C. Liu, J. Kramer, G. Indiveri, T. Delbru? ck, and R. Douglas, ?Orientation-selective aVLSI
spiking neurons,? in Advances in Neural Information Processing Systems 14 (T. G. Dietterich,
S. Becker, and Z. Ghahramani, eds.), Cambridge, MA: MIT Press, 2002.
[8] G. Indiveri, A. M. Whatley, and J. Kramer, ?A reconfigurable neuromorphic VLSI multi-chip
system applied to visual motion computation,? in Proc. MicroNeuro?99, Apr. 1999.
[9] D. H. Goldberg, G. Cauwenberghs, and A. G. Andreou, ?Probabilistic synaptic weighting in a
reconfigurable network of VLSI integrate-and-fire neurons,? Neural Networks, vol. 14, no. 6-7,
pp. 781?793, 2001.
(a)
(b)
Figure 7: (a) Stage 3 foveation network. The 32 ? 32 pixel high-resolution fovea (center)
is surrounded by lower-resolution areas where 2 ? 2, 4 ? 4, and 8 ? 8 blocks of OR
neurons (shown as non-overlapping for clarity) project to single IFAT cells. The address
space for inputs to the foveation network is 128 ? 128. [18]. (b) Output of stage 3, as
implemented on the IFAT, with the fovea centered on the location with the maximum firing
rate in Figure 6b, from stage 2. Peripheral pixels that receive no input are not shown.
[10] S. R. Deiss, R. J. Douglas, and A. M. Whatley, ?A pulse-coded communications infrastructure
for neuromorphic systems,? in Pulsed Neural Networks (W. Maass and C. M. Bishop, eds.),
pp. 157?178, Cambridge, MA: MIT Press, 1999.
[11] M. Mahowald and R. Douglas, ?A silicon neuron,? Nature, vol. 354, pp. 515?518, 1991.
[12] R. J. Vogelstein, F. Tenore, R. Philipp, M. S. Adlerstein, D. H. Goldberg, and G. Cauwenberghs,
?Spike timing-dependent plasticity in the address domain,? in Advances in Neural Information
Processing Systems 15 (S. Becker, S. Thrun, and K. Obermayer, eds.), Cambridge, MA: MIT
Press, 2003.
[13] R. J. Vogelstein, U. Mallik, and G. Cauwenberghs, ?Silicon spike-based synaptic array and
address-event transceiver,? in Proc. ISCAS?04, vol. 5, (Vancouver, BC), pp. 385?388, 2004.
[14] C. Koch, Biophysics of Computation: Information Processing in Single Neurons. New York,
NY: Oxford University Press, 1999.
[15] C. Koch and S. Ullman, ?Shifts in selective visual attention: towards the underlying neural
circuitry,? Human Neurobiology, vol. 4, pp. 219?227, 1985.
[16] L. Itti, E. Niebur, and C. Koch, ?A model of saliency-based fast visual attention for rapid scene
analysis,? IEEE Trans. Pattern Analysis & Machine Intelligence, vol. 20, no. 11, pp. 1254?1259,
1998.
[17] T. Horiuchi, T. Morris, C. Koch, and S. P. DeWeerth, ?Analog VLSI circuits for attentionbased, visual tracking,? in Advances in Neural Information Processing Systems 9, pp. 706?712,
Cambridge, MA: MIT Press, 1997.
[18] R. J. Vogelstein, U. Mallik, E. Culurciello, G. Cauwenberghs, and R. Etienne-Cummings, ?Spatial acuity modulation of an address-event imager,? in ICECS?04, 2004.
[19] R. J. Vogelstein, U. Mallik, and G. Cauwenberghs, ?Reconfigurable silicon array of spiking
neurons,? IEEE Trans. Neural Networks, 2005. (Submitted).
[20] E. Culurciello, R. Etienne-Cummings, and K. Boahen, ?Second generation of high dynamic
range, arbitrated digital imager,? in Proc. ISCAS?04, vol. 4, (Vancouver, BC), pp. 828?831,
2004.
| 2694 |@word version:1 c0:2 pulse:1 jacob:1 brightness:1 solid:1 reduction:1 liu:1 contains:3 series:1 bc:2 john:2 plasticity:1 shape:1 designed:6 intelligence:1 device:6 sram:1 realism:1 infrastructure:2 provides:2 location:4 philipp:1 constructed:1 c2:2 become:1 m7:1 transceiver:3 fabricate:1 inside:1 inter:2 rapid:3 frequently:1 multi:6 brain:1 m8:1 detects:1 initiating:1 little:1 provided:2 project:3 underlying:2 circuit:4 vthresh:1 mcu:4 sivilotti:2 cm:1 kind:1 developed:1 lfsr:2 pseudo:2 temporal:1 every:5 ti:2 imager:8 control:3 configuring:2 internally:1 unit:1 producing:1 before:1 service:1 engineering:4 local:6 timing:1 oxford:1 firing:3 modulation:4 dynamically:2 specifying:2 shaded:1 limited:3 projective:2 range:2 graduate:1 directed:1 acknowledgment:1 block:8 implement:10 differs:1 communicated:1 area:2 jhu:1 printed:1 convenient:1 word:2 integrating:1 cannot:1 demonstrated:2 center:5 attention:3 resolution:7 m2:1 higgins:1 array:15 creq:1 gert:2 increment:1 target:6 us:2 goldberg:2 hypothesis:1 element:2 npq:1 conserve:1 muri:1 bottom:1 m14:1 electrical:2 region:11 highest:3 observes:1 boahen:3 environment:1 dynamic:1 vdd:3 smart:1 exposed:2 triangle:1 rightward:1 resolved:1 darpa:1 chip:21 emulate:3 neurotransmitter:3 represented:1 surrounding:2 train:1 horiuchi:1 fast:3 describe:1 outside:1 plausible:1 reconstruct:1 compressed:1 final:1 whatley:2 relocate:1 mb:4 reset:1 neighboring:2 flexibility:1 description:1 normalize:1 los:1 transmission:1 extending:1 produce:1 advisable:1 recurrent:3 avlsi:4 received:1 implemented:9 come:1 direction:1 concentrate:1 functionality:4 centered:3 human:2 routing:2 enable:1 viewing:1 virtual:3 require:1 dio:4 ifat:33 subdivided:1 integrateand:1 biological:4 memorybased:1 around:1 koch:5 equilibrium:5 mapping:2 circuitry:3 driving:3 purpose:3 proc:4 vbp:1 sensitive:1 largest:1 organ:1 mit:4 sensor:5 always:1 activates:1 gaussian:1 modified:2 ck:1 varying:3 surveillance:2 voltage:1 release:1 acuity:5 indiveri:2 culurciello:4 contrast:5 dependent:1 stopping:1 typically:1 pasadena:1 vlsi:6 selective:2 ralph:1 pixel:11 orientation:1 spatial:6 field:9 having:1 identical:1 placing:1 look:2 cancel:1 throughput:1 future:1 haven:1 few:1 retina:2 surroundings:1 simultaneously:1 m4:1 phase:1 iscas:2 fire:9 versatility:1 arbitrated:1 conductance:2 interest:3 custom:3 light:9 necessary:1 circle:2 column:3 mhz:1 neuromorphic:8 mahowald:3 entry:1 uniform:1 fpga:6 stored:4 connect:1 m10:1 probabilistic:1 vm:2 receiving:1 pool:2 hopkins:2 attentionbased:1 transmitting:1 connectivity:4 w1:1 central:1 thesis:1 external:9 derivative:1 itti:1 ullman:1 potential:10 m9:1 converted:1 nonoverlapping:1 pooled:1 includes:2 configured:1 register:1 stream:1 performed:1 lot:1 microcontroller:2 cauwenberghs:5 start:1 parallel:2 capability:1 square:1 reserved:1 judgment:1 saliency:3 identify:2 vbn:1 niebur:1 monitoring:1 ago:1 processor:1 submitted:1 explain:1 synapsis:5 strongest:1 synaptic:16 ed:4 energy:1 frequency:1 pp:11 auditory:1 appends:1 positioned:1 feed:2 bidirectional:1 originally:1 cummings:3 response:2 synapse:7 arranged:1 execute:1 box:1 generality:1 biomedical:1 stage:31 deweerth:2 outgoing:3 receives:2 horizontal:1 dac:6 overlapping:3 rack:1 mode:2 gray:1 effect:1 dietterich:1 contain:2 maass:1 white:2 wiring:1 during:1 xilinx:1 m5:1 complete:2 demonstrate:2 performs:1 motion:2 interface:1 image:10 consideration:1 novel:1 common:1 specialized:1 spiking:9 physical:1 million:1 analog:7 m1:1 kluwer:1 silicon:10 cambridge:4 language:1 funded:1 moving:1 access:1 operating:1 base:1 add:1 perspective:1 irrelevant:1 driven:1 pulsed:1 manipulation:1 route:1 store:1 n00014:2 binary:2 arbitrarily:1 onr:2 transmitted:1 additional:1 dashed:1 ii:2 resolving:1 multiple:2 desirable:1 vogelstein:4 exceeds:1 academic:1 award:2 coded:1 controlled:1 biophysics:1 ae:5 vision:2 physically:1 represent:3 kernel:2 achieved:2 cell:13 c1:1 receive:4 addition:1 background:1 fellowship:1 baltimore:2 diagram:2 sends:1 publisher:1 w2:2 unlike:1 pass:1 sent:3 capacitor:1 m6:1 architecture:2 topology:3 converter:1 bandwidth:4 reduce:3 regarding:1 idea:1 identified:1 intensive:1 shift:3 motivated:1 utility:1 becker:2 stereo:1 reconfigurability:1 routed:1 york:1 cause:1 lazzaro:1 useful:2 amount:1 dark:2 morris:1 hardware:5 generate:2 nsf:2 inhibitory:5 stereoscopic:1 per:3 discrete:1 vol:9 four:6 salient:1 threshold:2 acknowledged:1 drawn:1 micrograph:2 clarity:3 douglas:3 ram:9 angle:3 communicate:1 respond:1 bit:6 entirely:1 ct:1 convergent:1 yale:2 aer:7 strength:1 x2:1 software:1 multiplexing:1 unlimited:1 scene:2 speed:1 span:1 performing:2 pseudorandom:1 according:3 peripheral:3 request:2 membrane:3 postsynaptic:5 wawrzynek:1 biologically:1 projecting:1 equation:1 bus:6 pin:1 addr:5 sending:1 operation:2 appropriate:4 appending:1 subtracted:1 original:1 top:1 etienne:4 ghahramani:1 classical:1 society:1 added:1 quantity:1 spike:15 alamitos:1 strategy:1 primary:1 receptive:2 md:2 icecs:1 cid:3 obermayer:1 fovea:9 separate:1 thrun:1 capacity:2 decoder:2 w0:2 originate:1 presynaptic:1 collected:2 code:1 quantal:1 executed:1 trace:2 design:3 implementation:2 proper:1 perform:2 m11:1 m12:1 neuron:57 vertical:1 vcomp:1 extended:1 communication:6 emulating:1 neurobiology:1 arbitrary:3 intensity:4 required:2 extensive:1 connection:9 deiss:1 conflict:2 andreou:1 instructing:1 california:1 accepts:2 alternately:1 trans:4 address:30 beyond:1 pattern:2 prototyping:2 rf:3 memory:4 video:1 power:1 event:28 overlap:1 serially:1 gillespie:1 hardwired:2 advanced:1 m15:1 scheme:2 technology:1 picture:1 identifies:1 eugenio:2 extract:1 vreset:1 literature:1 vancouver:2 generation:2 mounted:1 proportional:1 allocation:1 microneuro:1 digital:8 integrate:8 sufficient:1 port:1 surrounded:1 row:3 anniversary:1 course:1 excitatory:5 repeat:1 placed:1 asynchronous:4 supported:1 salience:9 allow:1 institute:1 wide:3 fall:1 biomorphic:1 outermost:1 feedback:1 dimension:1 transition:1 computes:2 sensory:2 forward:2 made:1 active:1 incoming:10 grayscale:1 search:1 decade:1 table:1 additionally:1 channel:1 nature:1 ca:2 m13:1 complex:1 protocol:1 impinging:1 octopus:2 domain:2 apr:1 identifier:1 x1:1 enlarged:1 site:2 board:2 nonvolatile:1 ny:1 decoded:1 candidate:1 communicates:2 third:2 weighting:1 externally:1 specific:1 reconfigurable:9 bishop:1 offset:4 divergent:1 weakest:1 modulates:1 phd:1 illustrates:1 downward:1 boston:1 simply:1 visual:9 tracking:1 partially:1 corresponds:1 ma:5 goal:1 marked:1 sized:1 kramer:2 towards:1 absence:1 content:1 change:4 foveation:4 reducing:1 operates:1 total:2 called:1 experimental:1 m3:1 selectively:3 internal:6 modulated:1 handshaking:3 dept:3 arbitration:5 |
1,866 | 2,695 | Breaking SVM Complexity
with Cross-Training
G?okhan H. Bak?r
Max Planck Institute
for Biological Cybernetics,
T?ubingen, Germany
[email protected]
L?eon Bottou
NEC Labs America
Princeton NJ, USA
[email protected]
Jason Weston
NEC Labs America
Princeton NJ, USA
[email protected]
Abstract
We propose to selectively remove examples from the training set using
probabilistic estimates related to editing algorithms (Devijver and Kittler,
1982). This heuristic procedure aims at creating a separable distribution
of training examples with minimal impact on the position of the decision
boundary. It breaks the linear dependency between the number of SVs
and the number of training examples, and sharply reduces the complexity
of SVMs during both the training and prediction stages.
1
Introduction
The number of Support Vectors (SVs) has a dramatic impact on the efficiency of Support
Vector Machines (Vapnik, 1995) during both the learning and prediction stages. Recent
results (Steinwart, 2004) indicate that the number k of SVs increases linearly with the
number n of training examples. More specifically,
k/n ?? 2BK
(1)
where n is the number of training examples and BK is the smallest classification error
achievable with the SVM kernel K. When using a universal kernel such as the Radial Basis
Function kernel, BK is the Bayes risk B, i.e. the smallest classification error achievable with
any decision function.
The computational requirements of modern SVM training algorithms (Joachims, 1999;
Chang and Lin, 2001) are very largely determined by the amount of memory required to
store the active segment of the kernel matrix. When this amount exceeds the available
memory, the training time increases quickly because some kernel matrix coefficients must
be recomputed multiple times. During the final phase of the training process, the active
segment always contains all the k(k + 1)/2 dot products between SVs. Steinwart?s result
(1) then suggests that the critical amount of memory scales at least like B 2 n2 . This can be
practically prohibitive for problems with either big training sets or large Bayes risk (noisy
problems). Large numbers of SVs also penalize SVMs during the prediction stage as the
computation of the decision function requires a time proportional to the number of SVs.
When the problem is separable, i.e. B = 0, equation (1) suggests1 that the number k of
SVs increases less than linearly with the number n of examples. This improves the scaling
laws for the SVM computational requirements.
1
See also (Steinwart, 2004, remark 3.8)
In this paper, we propose to selectively remove examples from the training set using probabilistic estimates inspired by training set editing algorithms (Devijver and Kittler, 1982).
The removal procedure aims at creating a separable set of training examples without modifying the location of the decision boundary. Making the problem separable breaks the
linear dependency between the number of SVs and the number of training examples.
2
2.1
Related work
Salient facts about SVMs
We focus now on the C-SVM applied to the two-class pattern recognition problem. See
(Burges, 1998) for a concise reference. Given n training patterns xi and their associated
classes yi = ?1, the SVM decision function is:
f (x) =
n
X
(2)
?i? yi K(xi , x) + b?
i=1
The coefficient ?i? in (2) are obtained by solving a quadratic programing problem:
??
=
1X
?i ?j yi yj K(xi , xj )
2 i,j
?
i
X
subject to ?i, 0 ? ?i ? C and
?i yi = 0
arg max
X
?i ?
(3)
i
This optimization yields three categories of training examples depending on ? i? . Within
each category, the possible values of the margins yi f (xi ) are prescribed by the KarushKuhn-Tucker optimality conditions.
- Examples such that ?i? = C are called bouncing SVs or margin errors and satisfy
yi f (xi ) < 1. The set of bouncing SVs includes all training examples misclassified
by the SVM, i.e. those which have a negative margin yi f (xi ) < 0.
- Examples such that 0 < ?i? < C are called ordinary SVs and satisfy yi f (xi ) = 1.
- Examples such that ?i? = 0 satisfy relation yi f (xi ) > 1. These examples play no
role in the SVM decision function (2). Retraining after discarding these examples
would still yield the same SVM decision function (2).
These facts provide some insight into Steinwart?s result (1). The SVM decision function,
like any other decision rule, must asymptotically misclassify at least Bn examples, where
B is the Bayes risk. All these examples must therefore become bouncing SVs.
To illustrate dependence on the Bayes risk, we perform a linear classification task in two
dimensions under varying amount of class overlap. The class distributions were uniform
on a unit square with centers c1 and c2 . Varying the distance between c1 and c2 allows us
to control the Bayes risk. The results are shown in figure 1.
2.2
A posteriori reduction of the number of SVs.
Several techniques aim to reduce the prediction complexity of SVMs by expressing the
SVM solution (2) with a smaller kernel expansion. Since one must compute the SVM solution before applying these post-processing techniques, they are not suitable for reducing
the complexity of the training stage.
Reduced Set Construction. Burges (Burges, 1996) proposes to construct new patterns
zj in order to define a compact approximation of the decision function (2). Reduced set
construction usually involves solving a non convex optimization problem and is not applicable on arbitrary inputs such as graphs or strings.
3
10
# ?=C for SVM
# ?<C for SVM
rank K
sv
1.5
y.f(x)
1
2
10
log #SV
0.5
?
0
?0.5
1
10
?1
?1.5
1000 2000 3000 4000 5000 6000 7000
0
10
0
0.05
0.1
0.15
0.2
Bayes risk
0.25
0.3
0.35
Figure 1: Effect of noise on the number of sup-
Figure 2: Histogram of SVs selected by the
port vectors. The number of ordinary SVs stays
almost constant whereas the number of bouncing SVs grows. Additional support vectors do
not give extra information as indicated by the
rank of the kernel matrix. See section 2.1.
`1 penalization method on the MNIST 3-8 discrimination task. The initial SVs have been ordered on the x-axis by increasing margin yf (x)
and decreasing ?. See last paragraph in section
2.2.
Reduced Set Selection. The set of basis2 functions K(xi , ?) associated with the SVs xi
do not necessarily constitute a linearly independent family. The same decision function f (?)
can then be expressed by multiple linear combination of the functions K(xi , ?). Reduced
set selection methods attempt to select a subset of the SVs that is sufficient to express
the SVM decision function. For instance, (Downs, Gates and Masters, 2001) propose to
compute the row echelon form of the kernel matrix and discard SVs that lead to zero rows.
This approach maintains the original SVM decision function.
In contrast, the `1 penalization method suggested in (Sch?olkopf and Smola, 2002,
sect. 18.4.2) simply attempts to construct a sufficiently good approximation of the original SVM decision function by solving
2
X
X
X
?
arg min
?i yi K(xi , ?) ?
?i yi K(xi , ?)
+ ?
|?i |
(4)
?
i
i
K
i
where parameter ? trades accuracy versus sparsity, and k?kK denotes the Reproducing Kernel Hilbert Space norm (Sch?olkopf and Smola, 2002, definition 2.9). Simplifying expression (4) yields a numerically tractable quadratic programming problem.
Which examples are selected? We have investigated the `1 penalization method (4) as
follows. We train a first SVM to discriminate digits 3 and 8 on the MNIST dataset (see
section 4.2) after randomly swapping 10% of the class labels in the training set. We then
select a subset of the resulting support vectors using the `1 penalization method.
Choosing ? is quite difficult in practice. To evaluate the accuracy of the procedure, we train
a second SVM on the selected vectors, compare its recognition accuracy with that of the
first SVM. This was best achieved by enforcing the constraint ?i ? 0 in (4) because the
second SVM cannot return an expansion with negative coefficients.
Figure 2 shows the histogram of selected SVs. The initial support vectors have been ordered on the x-axis by increasing values of yi f (xi ), and, in the case of margin SVs, by
decreasing values of ?i . The selected SVs includes virtually no misclassified SVs, but
instead concentrates on SVs with large ?i .
This result suggests that simple pre-processing methods might indicate which training examples are really critical for SVM classification.
2
We use the customary name basis functions despite linear dependence. . .
2.3
Training set editing techniques
We now consider techniques for reducing the set of training examples before running a
training algorithm. Reducing the amount of training data is indeed an obvious way to
reduce the complexity of training. Quantization and clustering methods might be used to
achieve this goal. These methods however reduce the training data without considering the
loss function of interest, and therefore sacrifice classification accuracy. We focus instead
on editing techniques, i.e. techniques for discarding selected training examples with the
aim of achieving similar or better classification accuracy.
Two prototypical editing techniques, MULTIEDIT and CONDENSE, have been thoroughly
studied (Devijver and Kittler, 1982, chapter 3) in the context of the nearest neighbor (1NN) classification rule.
Removing interior examples. The CONDENSE algorithm was first described by (Hart,
1968). This algorithm selects a subset of the training examples whose 1-NN decision
boundary still classifies correctly all of the initial training examples:
Algorithm 1 (C ONDENSE).
1 Select a random training example and put it in set R.
2 For each training example i = 1, . . . , n : classify example i using the 1-NN rule with set
R as the training set, and insert it into R if it is misclassified.
3 Return to step 2 if R has been modified during the last pass.
4 The final contents of R constitute the condensed training set.
This is best understood when both classes form homogeneous clusters in the feature space.
Algorithm 1 discards training examples located in the interior of each cluster.
This strategy works poorly when there is a large overlap between the pattern distributions
of both classes, that is to say when the Bayes risk B is large. Consider for instance a feature
space region where P (y = +1 | x) > P (y = ?1 | x) > 0. A small number of training
examples of class y = ?1 can still appear in such a region. We say that they are located on
the wrong side of the Bayes decision boundary. Asymptotically, all such training examples
belong to the condensed training set in order to ensure that they are properly recognized as
members of class y = ?1.
Removing noise examples. The Edited Nearest Neighbor rule (Wilson, 1972) suggests
to first discard all training examples that are misclassified when applying the 1-NN rule
using all n ? 1 remaining examples as the training set. It was shown that removing these
examples improves the asymptotic performance of the nearest neighbor rule. Whereas
the 1-NN risk is asymptotically bounded by 2B, the Edited 1-NN risk is asymptotically
bounded by 1.2 B, where B is the Bayes risk.
The MULTIEDIT algorithm (Devijver and Kittler, 1982, section 3.11) asymptotically discards all the training examples located on the wrong side of the Bayes decision boundary.
The asymptotic risk of the multi-edited nearest neighbor rule is the Bayes risk B.
Algorithm 2 (M ULTI E DIT).
1 Divide randomly the training data into s splits S1 , . . . , Ss . Let us call fi the 1-NN classifier
that uses Si as the training set.
2 Classify all examples in Si using the classifier f(i+1) mod s and discard all misclassified
examples.
3 Gather all the remaining examples and return to step 1 if any example has been discarded
during the last T iterations.
4 The remaining examples constitute the multiedited training set.
By discarding examples located on the wrong side of the Bayes decision boundary, algorithm MULTIEDIT constructs a new training set whose apparent distribution has the same
Bayes decision boundary as the original problem, but with Bayes risk equal to 0. Devijver
and Kittler claim that MULTIEDIT produces an ideal training set for CONDENSE.
Algorithm MULTIEDIT also discards some proportion of training examples located on the
correct side of Bayes decision boundary. Asymptotically this does not matter. However
this is often a problem in practice. . .
2.4
Editing algorithms and SVMs
Training examples recognized with high confidence usually do not appear in the SVM
solution (2) because they do not become support vectors. On the other hand, outliers always
become support vectors. Intuitively, SVMs display the properties of the CONDENSE but
lack the properties of the MULTIEDIT algorithm.
The mathematical proofs for the asymptotic properties of MULTIEDIT depend on the specific nature of the 1-NN classification rule. The MULTIEDIT algorithm itself could be identically defined for any classifier. This suggests (but does not prove) that these properties
might remain valid for SVM classifiers3 .
This contribution is an empirical attempt to endow Support Vector Machines with the properties of the MULTIEDIT algorithm.
Editing SVM training sets implicitly modifies the SVM loss function in a way that relates
to robust statistics. Editing alters the apparent distribution of training examples such that
the class distributions P (x | y = 1) and P (x | y = ?1) no longer overlap. If the class
distributions were known, this could be done by trimming the tails of the class distributions.
A similar effect could be obtained by altering the SVM loss function (the hinge loss) into a
non convex loss function that gives less weight to outliers.
3
Cross-Training
Cross-Training is a representative algorithm of such combinations of SVMs and editing
algorithms. It begins with creating s subsets of the training set with r examples each. Independent SVMs are then trained on each subset. The decision functions of these SVMs
are then used to discard two types of training examples: those which are confidently recognized, as in CONDENSE, and those which are misclassified, as in MULTIEDIT. A final SVM
is then trained using the remaining examples.
Algorithm 3 (C ROSS T RAINING).
1 Create s subsets of size r by randomly drawing r/2 examples of each class.
2 Train s independent SVMs f1 , . . . , fs using each of the subsets as the training set.
3 For each training example (xi , yi ) estimate the margin average mi and variance vi :
P
P
mi = 1s sr=1 yi fr (xi )
vi = 1s sr=1 (mi ? yi fr (xi ))2
4 Discard all training examples for which mi + vi < 0.
5 Discard all training examples for which mi ? vi > 1.
6 Train a final SVM on the remaining training examples.
The apparent simplicity of this algorithm hides a lot of hyperparameters. The value of the
C parameters for the SVMs at steps [2] and [6] has a considerable effect on the overall
performance of the algorithm.
For the first stage SVMs, we choose the C parameter which yields the best performance
on training sets of size r. For the second stage SVMs, we choose the C parameter which
yields the best overall performance measured on a separate validation set.
Furthermore, we discovered that the discarding steps tend to produce a final set of training
examples with very different numbers of examples for each class. Specific measures to
alleviate this problem are discussed in section 4.3.
3
Further comfort comes from the knowledge that a SVM with the RBF kernel and without threshold term b implements the 1-NN rule when the RBF radius tends to zero.
0.2
LIBSVM
X?Train LIBSVM
0.19
2500
500
LIBSVM
X?Train LIBSVM
450
400
0.18
2000
350
300
SVs
Test Error
0.17
1500
250
0.16
1000
200
150
0.15
100
500
0
0
LIBSVM
X?Train LIBSVM
Time (secs)
3000
0.14
1000
2000
3000 4000 5000
Training set size
6000
7000
8000
0.13
0
50
1000
2000
3000 4000 5000
Training set size
6000
7000
8000
0
0
1000
2000
3000 4000 5000
Training set size
6000
7000
8000
Figure 3: Comparing LIBSVM and Cross-Training on a toy problem of two Gaussian clouds for
increasing number of training points. Cross-Training gives an almost constant number of support
vectors (left figure) for increasing training set size, whereas in LIBSVM the number of support vectors increases linearly. The error rates behave similarly (middle figure), and Cross-Training gives an
improved training time (right figure). See section 4.1.
4
Experiments
4.1
Artificial Data
We first constructed artificial data, by generating two classes from two Gaussian clouds in
10 dimensions with means (1, 1, 1, 1, 1, 0, 0, 0, 0, 0) and (?1, ?1, ?1, ?1, ?1, 0, 0, 0, 0, 0)
and standard deviation 4. We trained a linear SVM for differing amounts of training points,
selecting C via cross validation. We compare the performance of LIBSVM4 with CrossTraining using LIBSVM with s = 5, averaging over 10 splits. The results given in figure
3 show a reduction in SVs and computation time using Cross-Training, with no loss in
accuracy.
4.2
Artificial Noise
Our second experiment involves the discrimination of digits 3 and 8 in the MNIST5
database. Artificial noise was introduced by swapping the labels of 0%, 5%, 10% and
15% of the examples. There are 11982 training examples and 1984 testing examples. All
experiments were carried out using LIBSVM?s ?-SVM (Chang and Lin, 2001) with the
RBF kernel (? = 0.005). Cross-Training was carried out by splitting the 11982 training
examples into 5 subsets. Figure 4 reports our results for various amounts of label noise.
The number of SVs (left figure) increases linearly for the standard SVM and stays constant
for the Cross-Training SVM. The test errors (middle figure) seem similar. Since our label
noise is artificial, we can also measure the misclassification rate on the unmodified testing set (right figure). This measurement shows a slight loss of accuracy without statistical
significance.
4.3
Benchmark Data
Finally the cross-training algorithm was applied to real data sets from both the ANU repository6 and from the the UCI repository7 .
Experimental results were quite disappointing until we realized that the discarding steps
tends to produce training sets with very different numbers of examples for each class. To
alleviate this problem, after training each SVM, we choose the value of the threshold b ?
4
5
6
7
http://www.csie.ntu.edu.tw/?cjlin/libsvm/
http://yann.lecun.com/exdb/mnist
http://mlg.anu.edu.au/?raetsch/data/index.html
ftp://ftp.ics.uci.edu/pub/machine-learning-databases
8000
2.5%
15.0%
2.0%
6000
1.5%
10.0%
4000
1.0%
5.0%
2000
0.5%
0
0.0%
0%
5%
10%
15%
0.0%
0%
5%
10%
15%
0%
5%
10%
15%
Figure 4: Number of SVs (left figure) and test error (middle figure) for varying amounts of label
noise on the MNIST 3-8 discrimination task. The x-axis in all graphs shows the amount of label noise;
white squares correspond to LIBSVM; black circles to Cross-Training; dashed lines to bagging the
first stage Cross-Training SVMs. The last graph (right figure) shows the test error measured without
label noise. See section 4.2
in (2) which achieves the best validation performance. We also attempt to balance the
final training set by re-inserting examples discarded during step [5] of the cross-training
algorithm.
Experiments were carried out using RBF kernels with the kernel width reported in the
literature. In the SVM experiments, the value of parameter C was determined by crossvalidation and then used for training a SVM on the full dataset. In the cross-training experiments, we make a validation set by taking r/3 examples from the training set. These
examples are only used for choosing the values of C and for adjusting the SVM thresholds.
Details and source code are available8 .
Dataset
Banana
Waveform
Splice
Adult
Adult
Forest
Forest
Forest
Train
Size
400
400
1000
3185
32560
50000
90000
200000
Test
Size
4900
4600
2175
16280
16280
58100
58100
58100
SVM
Perf.[%]
89.0
90.2
90.0
84.2
85.1
90.3
91.6
?
SVM
#SV
111
172
601
1207
11325
12476
18983
?
XTrain
Subsets
5?200
5?200
5?300
5?700
5?6000
5?10000
5?18000
8?30000
XTrain
Perf.[%]
88.2
88.7
89.9
84.2
84.8
89.2
90.7
92.1
XTrain
#SV
51
87
522
606
1194
7967
13023
19526
Table 1: Comparison of SVM and Cross-Training results on standard benchmark data sets.
The columns in table 1 contain the dataset name, the size of the training set used for the
experiment, the size of the test set, the SVM accuracy and number of SVs, the CrossTraining subset configuration, accuracy, and final number of SVs. Bold typeface indicates
which differences were statistically significant according to a paired test. These numbers
should be considered carefully because they are impacted by the discrete nature of the grid
search for parameter C. The general trend still indicates that Cross-Training causes a slight
loss of accuracy but requires much less SVs.
Our largest training set contains 200000 examples. Training a standard SVM on such a
set takes about one week of computation. We do not report this result because it was not
practical to determine a good value of C for this experiment. Cross-Training with specified
hyperparameters runs overnight. Cross-Training with hyperparameter grid searches runs in
two days.
We do not report detailled timing results because much of the actual time can be attributed
8
http://www.kyb.tuebingen.mpg.de/bs/people/gb/xtraining
to the search for the proper hyperparameters. Timing results would then depend on loosely
controlled details of the hyperparameter grid search algorithms.
5
Discussion
We have suggested to combine SVMs and training set editing techniques to break the linear relationship between number of support vectors and number of examples. Such combinations raise interesting theoretical questions regarding the relative value of each of the
training examples.
Experiments with a representative algorithm, namely Cross-Training, confirm that both the
training and the recognition time are sharply reduced. On the other hand, Cross-Training
causes a minor loss of accuracy, comparable to that of reduced set methods (Burges, 1996),
and seems to be more sensitive than SVMs in terms of parameter tuning.
Despite these drawbacks, Cross-Training provides a practical means to construct kernel
classifiers with significantly larger training sets.
6
Acknowledgement
We thank Hans Peter Graf, Eric Cosatto and Vladimir Vapnik for their advice and support.
Part of this work was funded by NSF grant CCR-0325463.
References
Burges, C. J. C. (1996). Simplified Support Vector Decision Rules. In Saitta, L., editor, Proceedings of the 13th International Conference on Machine Learning, pages 71?77, San Mateo, CA.
Morgan Kaufmann.
Burges, C. J. C. (1998). A Tutorial on Support Vector Machines for Pattern Recognition. Data
Mining and Knowledge Discovery, 2(2):121?167.
Chang, C.-C. and Lin, C.-J. (2001). Training ?-Support Vector Classifiers: Theory and Algorithms.
Neural Computation, 13(9):2119?2147.
Devijver, P. and Kittler, J. (1982). Pattern Recogniton, A statistical approach. Prentice Hall, Englewood Cliffs.
Downs, T., Gates, K. E., and Masters, A. (2001). Exact Simplification of Support Vector Solutions.
Journal of Machine Learning Research, 2:293?297.
Hart, P. (1968). The condensed nearest neighbor rule. IEEE Transasctions on Information Theory,
14:515?516.
Joachims, T. (1999). Making Large?Scale SVM Learning Practical. In Sch o? lkopf, B., Burges, C.
J. C., and Smola, A. J., editors, Advances in Kernel Methods ? Support Vector Learning, pages
169?184, Cambridge, MA. MIT Press.
Sch?olkopf, B. and Smola, A. J. (2002). Learning with Kernels. MIT Press, Cambridge, MA.
Steinwart, I. (2004). Sparseness of Support Vector Machines?Some Asymptotically Sharp Bounds.
In Thrun, S., Saul, L., and Scho? lkopf, B., editors, Advances in Neural Information Processing
Systems 16. MIT Press, Cambridge, MA.
Vapnik, V. N. (1995). The Nature of Statistical Learning Theory. Springer Verlag, New York.
Wilson, D. L. (1972). Asymptotic properties of the nearest neighbor rules using edited data. IEEE
Transactions on Systems, Man, and Cybernetics, 2:408?420.
| 2695 |@word middle:3 achievable:2 norm:1 proportion:1 retraining:1 seems:1 bn:1 simplifying:1 concise:1 dramatic:1 multiedit:10 reduction:2 initial:3 configuration:1 contains:2 selecting:1 pub:1 com:2 comparing:1 si:2 must:4 kyb:1 remove:2 discrimination:3 prohibitive:1 selected:6 provides:1 location:1 org:1 mathematical:1 c2:2 constructed:1 become:3 prove:1 combine:1 paragraph:1 sacrifice:1 indeed:1 mpg:2 multi:1 inspired:1 decreasing:2 actual:1 considering:1 increasing:4 begin:1 classifies:1 bounded:2 string:1 differing:1 nj:2 wrong:3 classifier:5 control:1 unit:1 grant:1 appear:2 planck:1 before:2 understood:1 timing:2 tends:2 despite:2 cliff:1 might:3 black:1 au:1 studied:1 mateo:1 suggests:4 statistically:1 practical:3 lecun:1 yj:1 testing:2 practice:2 implement:1 digit:2 procedure:3 empirical:1 universal:1 significantly:1 pre:1 radial:1 confidence:1 cannot:1 interior:2 selection:2 put:1 risk:13 applying:2 context:1 prentice:1 www:2 center:1 modifies:1 convex:2 simplicity:1 splitting:1 insight:1 rule:12 construction:2 play:1 exact:1 programming:1 homogeneous:1 us:1 trend:1 recognition:4 located:5 database:2 role:1 cloud:2 csie:1 kittler:6 region:2 sect:1 trade:1 edited:4 complexity:5 trained:3 depend:2 solving:3 segment:2 raise:1 efficiency:1 eric:1 basis:2 chapter:1 america:2 various:1 train:8 artificial:5 choosing:2 quite:2 heuristic:1 whose:2 apparent:3 larger:1 say:2 s:1 drawing:1 statistic:1 noisy:1 itself:1 final:7 propose:3 product:1 fr:2 inserting:1 uci:2 poorly:1 achieve:1 olkopf:3 crossvalidation:1 cluster:2 requirement:2 produce:3 generating:1 ftp:2 depending:1 illustrate:1 measured:2 nearest:6 minor:1 involves:2 indicate:2 come:1 overnight:1 bak:1 concentrate:1 radius:1 waveform:1 correct:1 drawback:1 modifying:1 f1:1 really:1 alleviate:2 ntu:1 biological:1 insert:1 practically:1 sufficiently:1 considered:1 ic:1 hall:1 week:1 claim:1 achieves:1 smallest:2 applicable:1 condensed:3 label:7 ross:1 sensitive:1 largest:1 create:1 mit:3 always:2 gaussian:2 aim:4 modified:1 varying:3 wilson:2 endow:1 focus:2 joachim:2 properly:1 rank:2 indicates:2 contrast:1 posteriori:1 saitta:1 nn:9 relation:1 misclassified:6 condense:5 germany:1 selects:1 arg:2 classification:8 overall:2 html:1 proposes:1 equal:1 construct:4 report:3 modern:1 randomly:3 phase:1 attempt:4 misclassify:1 interest:1 trimming:1 englewood:1 mining:1 swapping:2 cosatto:1 divide:1 loosely:1 circle:1 re:1 theoretical:1 minimal:1 instance:2 classify:2 column:1 altering:1 unmodified:1 ordinary:2 deviation:1 subset:10 uniform:1 reported:1 dependency:2 sv:4 thoroughly:1 international:1 stay:2 probabilistic:2 quickly:1 choose:3 creating:3 return:3 toy:1 de:2 sec:1 bold:1 includes:2 coefficient:3 matter:1 satisfy:3 vi:4 break:3 jason:1 lab:3 lot:1 sup:1 bayes:15 maintains:1 contribution:1 square:2 accuracy:11 variance:1 largely:1 kaufmann:1 yield:5 correspond:1 lkopf:2 cybernetics:2 definition:1 mlg:1 tucker:1 obvious:1 associated:2 proof:1 mi:5 attributed:1 dataset:4 adjusting:1 knowledge:2 improves:2 hilbert:1 carefully:1 day:1 impacted:1 improved:1 editing:10 done:1 typeface:1 furthermore:1 stage:7 smola:4 until:1 hand:2 steinwart:5 lack:1 yf:1 indicated:1 grows:1 name:2 effect:3 usa:2 contain:1 white:1 during:7 width:1 ulti:1 exdb:1 fi:1 scho:1 belong:1 tail:1 discussed:1 slight:2 numerically:1 expressing:1 measurement:1 raetsch:1 significant:1 cambridge:3 tuning:1 grid:3 similarly:1 dot:1 funded:1 han:1 longer:1 recent:1 hide:1 discard:9 disappointing:1 store:1 verlag:1 ubingen:1 yi:15 morgan:1 additional:1 recognized:3 determine:1 dashed:1 relates:1 multiple:2 full:1 reduces:1 exceeds:1 libsvm4:1 cross:22 okhan:1 lin:3 hart:2 post:1 paired:1 controlled:1 impact:2 prediction:4 histogram:2 kernel:16 iteration:1 achieved:1 penalize:1 c1:2 whereas:3 source:1 sch:4 extra:1 sr:2 subject:1 tend:1 virtually:1 member:1 mod:1 seem:1 call:1 ideal:1 split:2 identically:1 xj:1 reduce:3 regarding:1 expression:1 gb:2 f:1 peter:1 york:1 cause:2 svs:32 remark:1 constitute:3 xtrain:3 amount:9 svms:16 category:2 dit:1 reduced:6 http:4 zj:1 nsf:1 alters:1 tutorial:1 correctly:1 ccr:1 discrete:1 hyperparameter:2 express:1 recomputed:1 salient:1 threshold:3 achieving:1 libsvm:12 asymptotically:7 graph:3 run:2 master:2 bouncing:4 almost:2 family:1 yann:1 decision:22 scaling:1 comparable:1 bound:1 simplification:1 display:1 quadratic:2 constraint:1 sharply:2 prescribed:1 optimality:1 leon:1 min:1 separable:4 according:1 combination:3 smaller:1 remain:1 tw:1 making:2 s1:1 b:1 outlier:2 intuitively:1 equation:1 cjlin:1 tractable:1 available:1 gate:2 customary:1 original:3 bagging:1 denotes:1 running:1 clustering:1 ensure:1 remaining:5 hinge:1 eon:1 question:1 realized:1 strategy:1 dependence:2 distance:1 separate:1 thank:1 thrun:1 tuebingen:2 enforcing:1 code:1 index:1 relationship:1 kk:1 balance:1 vladimir:1 difficult:1 negative:2 proper:1 perform:1 recogniton:1 discarded:2 benchmark:2 behave:1 banana:1 discovered:1 reproducing:1 arbitrary:1 sharp:1 bk:3 introduced:1 namely:1 required:1 specified:1 adult:2 suggested:2 usually:2 pattern:6 comfort:1 sparsity:1 confidently:1 max:2 memory:3 critical:2 overlap:3 suitable:1 misclassification:1 axis:3 carried:3 perf:2 literature:1 acknowledgement:1 removal:1 discovery:1 asymptotic:4 law:1 relative:1 loss:9 graf:1 prototypical:1 interesting:1 proportional:1 versus:1 penalization:4 validation:4 gather:1 sufficient:1 port:1 editor:3 row:2 last:4 side:4 burges:7 institute:1 neighbor:6 saul:1 taking:1 boundary:8 dimension:2 raining:1 valid:1 san:1 simplified:1 transaction:1 compact:1 implicitly:1 confirm:1 active:2 xi:17 search:4 table:2 nature:3 robust:1 ca:1 forest:3 expansion:2 bottou:2 necessarily:1 investigated:1 significance:1 linearly:5 big:1 noise:9 hyperparameters:3 n2:1 advice:1 representative:2 position:1 breaking:1 splice:1 down:2 removing:3 discarding:5 specific:2 svm:44 mnist:4 vapnik:3 quantization:1 nec:3 anu:2 margin:6 sparseness:1 simply:1 expressed:1 ordered:2 chang:3 springer:1 devijver:6 ma:3 weston:1 goal:1 rbf:4 man:1 content:1 considerable:1 programing:1 specifically:1 determined:2 reducing:3 averaging:1 called:2 discriminate:1 pas:1 experimental:1 selectively:2 select:3 support:18 people:1 evaluate:1 princeton:2 |
1,867 | 2,696 | Identifying protein-protein interaction sites on a
genome-wide scale
Haidong Wang? Eran Segalo Asa Ben-Hur? Daphne Koller? Douglas L. Brutlag?
?
Computer Science Department, Stanford University, CA 94305
{haidong, koller}@cs.stanford.edu
o
Center for Studies in Physics and Biology, Rockefeller University, NY 10021
[email protected]
?
Department of Genome Sciences, University of Washington, WA 98195
[email protected]
?
Department of Biochemistry, Stanford University, CA 94305
[email protected]
Abstract
Protein interactions typically arise from a physical interaction of one or
more small sites on the surface of the two proteins. Identifying these sites
is very important for drug and protein design. In this paper, we propose
a computational method based on probabilistic relational model that attempts to address this task using high-throughput protein interaction data
and a set of short sequence motifs. We learn the model using the EM
algorithm, with a branch-and-bound algorithm as an approximate inference for the E-step. Our method searches for motifs whose presence in a
pair of interacting proteins can explain their observed interaction. It also
tries to determine which motif pairs have high affinity, and can therefore
lead to an interaction. We show that our method is more accurate than
others at predicting new protein-protein interactions. More importantly,
by examining solved structures of protein complexes, we find that 2/3 of
the predicted active motifs correspond to actual interaction sites.
1
Introduction
Many cellular functions are carried out through physical interactions between proteins.
Discovering the protein interaction map can therefore help to better understand the workings of the cell. Indeed, there has been much work recently on developing high-throughput
methods to produce a more complete map of protein-protein interactions [1, 2, 3].
Interactions between two proteins arise from physical interactions between small regions on the surface of the proteins [4] (see Fig. 2(b)). Finding interaction sites is an
important task, which is of particular relevance to drug design. There is currently no highthroughput experimental method to achieve this goal, so computational methods are required. Existing methods either require solving a protein?s 3D structure (e.g., [5]), and
therefore are computationally very costly and not applicable on a genome-wide scale, or
use known interaction sites as training data (e.g., [6]), which are relatively scarce and hence
have poor coverage. Other work focuses on refining the highly noisy high-throughput interaction maps [7, 8, 9], or on assessing the confidence levels of the observed interactions [10].
In this paper, we propose a computational method for predicting protein interactions
P1
d
d
a
Aa
P5
P2
Ad
Ab
P5
Ad
Ad
P1
Aab
Bab S Bdb
b
d
c
b
P2
b
P3
(a)
Adb
P4
T12 O
Aad
Add
Bab S Bab
T15 O
Abd
Bab S
T25 O
(b)
Figure 1: (a) Simple illustration of our assumptions for protein-protein interactions. The small
elements denote motif occurrences on proteins, with red denoting active and gray denoting inactive
motifs. (b) A fragment of our probabilistic model, for the proteins P 1 , P2 , P5 . We use yellow to
denote an assignment of the value true, and black to denote the value false; full circles denote an
assignment observed in the data, and patterned circles an assignment hypothesized by our algorithm.
The dependencies involving inactive motif pairs were removed from the graph because they do not
affect the rest of the model.
and the sites at which the interactions take place, which uses as input only high-throughput
protein-protein interaction data and the protein sequences. In particular, our method assumes no knowledge of the 3D protein structure, or of the sites at which binding occurs.
Our approach is based on the assumption that interaction sites can be described using a
limited repertoire of conserved sequence motifs [11]. This is a reasonable assumption since
interaction sites are significantly more conserved than the rest of the protein surface [12].
Given a protein interaction map, our method tries to explain the observed interactions by
identifying a set of sites of motif occurrence on every pair of interacting proteins through
which the interaction is mediated. To understand the intuition behind our approach, consider the example of Fig. 1(a). Here, the interaction pattern of the protein P1 can best be
explained using the motif pair a, b, where a appears in P1 and b in the proteins P2 , P3 , P4
but not in P5 . By contrast, the motif pair d, b is not as good an explanation, because d
also appears in P5 , which has a different interaction pattern. In general, our method aims
to identify motif pairs that have high affinity, potentially leading to interaction between
protein pairs that contain them.
However, a sequence motif might be used for a different purpose, and not give rise to an
active binding site; it might also be buried inside the protein, and thus be inaccessible for
interaction. Thus, the appearance of an appropriate motif does not always imply interaction.
A key feature of our approach is that we allow each motif occurrence in a protein to be
either active or inactive. Interactions are then induced only by the interactions of highaffinity active motifs in the two proteins. Thus, in our example, the motif d in p2 is inactive,
and hence does not lead to an interaction between p2 and p4 , despite the affinity between
the motif pair c, d. We note that Deng et al. [8] proposed a somewhat related method for
genome-wide analysis of protein interaction data, based on protein domains. However,
their method is focused on predicting protein-protein interactions and not on revealing the
site of interaction, and they do not allow for the possibility that some domains are inactive.
Our goal is thus to identify two components: the affinities between pairs of motifs, and
the activity of the occurrences of motifs in different proteins. Our algorithm addresses this
problem by using the framework of Bayesian networks [13] and probabilistic relational
models [14], which allows us to handle the inherent noise in the protein interaction data
and the uncertain relationship between interactions and motif pairs. We construct a model
encoding our assumption that protein interactions are induced by the interactions of active
motif pairs. We then use the EM algorithm [15], to fill in the details of the model, learning
both the motif affinities and activities from the observed data of protein-protein interactions
and protein motif occurrences. We address the computational complexity of the E-step in
these large, densely connected models by using an approximate inference procedure based
on branch-and-bound.
We evaluated our model on protein-protein interactions in yeast and Prosite motifs [11].
As a basic performance measure, we evaluated the ability of our method to predict new
protein-protein interactions, showing that it achieves better performance than several other
models. In particular, our results validate our assumption that we can explain interactions
via the interactions of active sequence motifs. More importantly, we analyze the ability
of our method to discover the mechanism by which the interaction occurs. Finally, we
examined co-crystallized protein pairs where the 3D structure of the interaction is known,
so that we can determine the sites at which the interaction took place. We show that our
active motifs are more likely to participate in interactions.
2
The Probabilistic Model
The basic entities in our probabilistic model are the proteins and the set of sequence motifs
that can mediate protein interactions. Our model therefore contains a set of protein entities
P = {P1 , . . . , Pn }, with the motifs that occur in them. Each protein P is associated with
the set of motifs that occur in it, denoted by P.M . As we discussed, a key premise of
our approach is that a specific occurrence of a sequence motif may or may not be active.
Thus, each motif occurrence a ? P.M is associated with a binary-value variable P.A a ,
which takes the value true if Aa is active in protein P and false otherwise. We structure the
|
prior probability P (P.Aa = true) = min{0.8, 3+0.1?|P.M
}, to capture our intuition that
|P.M |
the number of active motifs in a protein is roughly a constant fraction of the total number
of motifs in the protein, but that even proteins with few motifs tend to have at least some
number of active motifs.
A pair of active motifs in two proteins can potentially bind and induce an interaction
between the corresponding proteins. Thus, in our model, a pair of proteins interact if each
contains an active motif, and this pair of motifs bind to each other. The probability with
which two motifs bind to each other is called their affinity. We encode this assumption by
including in our model entities Tij corresponding to a pair of proteins Pi , Pj . For each
pair of motifs a ? Pi .M and b ? Pj .M , we introduce a variable Tij .Aab , which is a
deterministic AND of the activity of these two motifs. Intuitively, this variable represents
whether the pair of motifs can potentially interact. The probability with which two active
motif occurrences bind is their affinity. We model the binding event between two motif
occurrences using a variable Tij .Bab , and define: P (Tij .Bab = true | Tij .Aab = true) =
?ab and P (Tij .Bab = true | Tij .Aab = false) = 0, where ?ab is the affinity between motifs
a and b. This model reflects our assumption that two motif occurrences can bind only if
they are both active, but their actual binding probability depends on their affinity. Note that
this affinity is a feature of the motif pair and does not depend on the proteins in which they
appear.
We must also account for interactions that are not explained by our set of motifs,
whether because of false positives in the data, or because of inadequacies of our model
or of our motif set. Thus, we add a spurious binding variable Tij .S, for cases where an
interaction between Pi and Pj exists, but cannot be explained well by our set of active
motifs. The probability that a spurious binding occurs is given by P (Tij .S = true) = ?S .
Finally, we observe an interaction between two proteins if and only if some form of
binding occurs, whether by a motif pair or a spurious binding. Thus, we define a variable
Tij .O, which represents whether protein i was observed to interact with protein j, to be
a deterministic OR of all the binding variables Tij .S and Tij .Bab . Overall, Tij .O is a
noisy-OR [13] of all motif pair variables Tij .Aab .
Note that our model accounts for both types of errors in the protein interaction data.
False negatives (missing interactions) in the data are addressed through the fact that the
presence of an active motif pair only implies that binding takes place with some probability.
False positives (wrong interactions) in the data are addressed through the introduction of
the spurious interaction variables.
The full model defines a joint probability distribution over the entire set of attributes:
Q Q
P (P.A,T.A,
Q T.B, T.S, T.O) = i a?Pi .M P (Pi .Aa )
Q
a?Pi .M,b?Pj .M P (Tij .Aab | Pi .Aa , Pj .Ab )P (Tij .Bab | Tij .Aab )
ij
P (Tij .S)P (Tij .O | Tij .B, Tij .S)
where each of these conditional probability distributions is as specified above. We use
? to denote the entire set of model parameters {?a,b }a,b ? {?S }. An instantiation of our
probabilistic model is illustrated in Fig. 1(b).
3
Learning the Model
We now turn to the task of learning the model from the data. In a typical setting, we are
given as input a protein interaction data set, specifying a set of proteins P and a set of
observed interacting pairs T.O. We are also given a set of potentially relevant motifs, and
the occurrences of these motifs in the different proteins in P. Thus, all the variables except
for the O variables are hidden. Our learning task is thus twofold: we need to infer the values
of the hidden variables, both the activity variables P.A, T.A, and the binding variables
T.B, T.S; we also need to find a setting of the model parameters ?, which specify the
motif affinities. We use a variant of the EM algorithm [15] to find both an assignment to
the parameters ?, and an assignment to the motif variables P.A, which is a local maximum
of the likelihood function P (T.O, P.A | ?). Note that, to maximize this objective, we
search for a MAP assignment to the motif activity variables, but sum out over the other
hidden variables. This design decision is reasonable in our setting, where determining motif
activities is an important goal; it is a key assumption for our computational procedure.
As in most applications of EM, our main difficulty arises in the E-step, where we need
to compute the distribution over the hidden variables given the settings of the observed
variables and the current parameter settings. In our model, any two motif variables (both
within the same protein and across different proteins) are correlated, as there exists a path
of influence between them in the underlying Bayesian network (see Fig. 1(c)). These correlations make the task of computing the posterior distribution over the hidden variables
intractable, and we must resort to an approximate computation. While we could apply a
general purpose approximate inference algorithm such as loopy belief propagation [16],
such methods may not converge in densely connected model such as this one, and there
are few guarantees on the quality of the results even if they do converge. Fortunately, our
model turns out to have additional structure that we can exploit. We now describe an approximate inference algorithm that is tailored to our model, and is guaranteed to converge
to a (strong) local maximum.
Our first observation is that the only variables that correlate the different protein pairs
Tij are the motif variables P.A. Given an assignment to these activity variables, the network decomposes into a set of independent subnetworks, one for each protein pair. Based
on this observation, we divide our computation of the E-step into two parts. In the first,
we find an assignment to the motif variables in each protein, P.A; in the second, we compute the posterior probability over the binding motif pair variables T.B, T.S, given the
assignment to the motif variables.
We begin by describing the second phase. We observe that, as all the motif pair variables, T.A, are fully determined by the motif variables, the only variables left to reason
about are the binding variables T.B and T.S. The variables for any pair Tij are independent of the rest of the model given the instantiation to T.A and the interaction evidence. That fact, combined with the noisy-OR form of the interaction, allows us to compute the posterior probability required in the E-step exactly and efficiently. Specifically,
the computation for the variables associated with a particular protein pair Tij is as follows, where we omit the common prefix Tij to simplify notation. If Aab = false, then
P (Bab = true | Aab = false, O, ?) = 0. Otherwise, if Aab = true, then
P (Bab = true | A, O, ?)
=
P (Bab | A, ?)P (O | Bab = true, A, ?)
.
P (O | A, ?)
The first term in the numerator is simply the motif affinity ?ab ; the second term is 1 if
O = true andQ0 otherwise. The numerator can easily be computed as P (O | A, ?) =
1 ? (1 ? ?S ) Aa,b =true (1 ? ?ab ). The computation for P (S) is very similar.
We now turn to the first phase, of finding a setting to all of the motif variables. Unfortunately, as we discussed, the model is highly interconnected, and a finding an optimal
joint setting to all of these variables P.A is intractable. We thus approximate finding this
joint assignment using a method that exploits our specific structure. Our method iterates
over proteins, finding in each iteration the optimal assignment to the motif variables of each
protein given the current assignment to the motif activities in the remaining proteins. The
process repeats, iterating over proteins, until convergence.
As we discussed, the likelihood of each assignment to Pi .A can be easily computed
using the method described above. However, the computation for each protein is still exponential in the number of motifs it contains, which can be large (e.g., 15). However, in
our specific model, we can apply the following branch-and-bound algorithm (similar to an
approach proposed by Henrion [17] for BN2O networks) to find the globally optimal assignment to the motif variables of each protein. The idea is that we search over the space
of possible assignments Pi .A for one that maximizes the objective we wish to maximize.
We can show that if making a motif active relative to one assignment does not improve the
objective, it will also not improve the objective relative to a large set of other assignments.
More precisely, let f (Pi .A) = P (Pi .A, P?i .A|O, ?) denote the objective we wish
to maximize, where P?i .A is the fixed assignment to motif variables in all proteins
except Pi . Let Pi .A?a denote the assignment to all the motif variables in Pi except
for Aa . We
Q compute the ratio of f after we switch Pi .Aa from false to true. Let
ha (Pj ) = Pj .Ab =true (1 ? ?ab ) denote the probability that motif a does not bind with
any active motif in Pj . We can now compute:
?a (Pi .A?a ) =
?
Y
1?j?n
Tij .O=false
f (Pi .Aa = true, Pi .A?a )
g
=
f (Pi .Aa = false, Pi .A?a )
1?g
Q
Y
1 ? (1 ? ?S )ha (Pj ) a6=b,Pi .Ab =true hb (Pj )
Q
ha (Pj ) ?
1 ? (1 ? ?S ) a6=b,Pi .Ab =true hb (Pj )
1?j?n
(1)
Tij .O=true
where g is the prior probability for a motif in protein Pi to be active.
Now, consider a different point in the search, where our current motif activity assignment is Pi .A0?a , which has all the active motifs in Pi .A?a and some additional ones. The
first two terms in the product of Eq. (1) are the same for ?a (Pi .A?a ) and ?a (Pi .A0?a ).
For the final term (the large fraction), one can show using some algebraic manipulation
that this term in ?a (Pi .A0?a ) is lower than that for ?a (Pi .A?a ). We conclude that
?a (Pi .A?a ) ? ?a (Pi .A0?a ), and hence that:
f (Pi .Aa = true, Pi .A0?a )
f (Pi .Aa = true, Pi .A?a )
?1 ?
? 1.
f (Pi .Aa = false, Pi .A?a )
f (Pi .Aa = false, Pi .A0?a )
It follows that, if switching motif a from inactive to active relative to Pi .A decreases f , it
will also decrease f if we have some additional active motifs.
We can exploit this property in a branch-and-bound algorithm in order to find the globally optimal assignment Pi .A. Our algorithm keeps a set V of viable candidates for motif
assignments. For presentation, we encode assignments via the set of active motifs they
contain. Initially, V contains only the empty assignment {}. We start out by considering
motif assignments with a single active motif. We put such an assignment {a} in V if its
f -score is higher than f ({}). Now, we consider assignments {a, b} that have two active
motifs. We consider {a, b} only if both {a} and {b} are in V . If so, we evaluate its f -score,
and add it to V if this score is greater than that of {a} and {b}. Otherwise, we throw it
away. We continue this process for all assignments of size k: For each assignment with
active motif set S, we test whether S ? {a} ? V for all a ? S; if we compare f (S) to each
f (S ? {a}), and add it if it dominates all of them. The algorithm terminates when, from
some k, no assignment of size k is saved.
To understand the intuition behind this pruning procedure, consider a candidate assignment {a, b, c, d}, and assume that {a, b, c} ? V , but {b, c, d} 6? V . In this case, we must
have that {b, c} ? V , but adding d to that assignment reduces the f -score. In this case, as
shown by our analysis, adding d to the superset {a, b, c} would also reduce the f -score.
This algorithm is still exponential in worst case. However, in our setting, a protein with
many motifs has a low prior probability that each of them is active. Hence, adding new
motifs is less likely to increase the f -score, and the algorithm tends to terminate quickly.
As we show in Section 4, this algorithm significantly reduces the cost of our procedure.
Our E-step finds an assignment to P.A which is a strong local optimum of the objective function max P (P.A | T.O, ?): The assignment has higher probability than any
assignment that changes any of the motif variables for any single protein. For that assignment, our algorithm also computes the distribution over all of the binding variables, as
described above. Using this completion, we can now easily compute the (expected) sufficient statistics for the different parameters in the model. As each of these parameters is a
simple binomial distribution, the maximum likelihood estimation in the M-step is entirely
standard; we omit details.
4
Results
We evaluated our model on reliable S. cerevisiae protein interactions data from MIPS [2]
and DIP [3] databases. As for non-interaction data, we randomly picked pairs of proteins
that have no common function and cellular location. This results in a dataset of 2275
proteins, 4838 interactions (Tij .O = true), and 9037 non-interactions (Tij .O = false). We
used sequence motifs from the Prosite database [11] resulting in a dataset of 516 different
motifs with an average of 7.1 motif occurrences per protein. If a motif pair doesn?t appear
between any pair of interacting proteins, we initialize its affinity to be 0 to maximize the
joint likelihood. Its affinity will stay at 0 during the EM iterations and thus simplify our
model structure. We set the initial affinity for the remaining 8475 motif pairs to 0.03.
We train our model with motifs initialized to be either all active (P.A = true) or all
inactive (P.A = false). We get similar results with these two different initializations,
indicating the robustness of our algorithm. Below we only report the results based on all
motifs initialized to be active. Our branch-and-bound algorithm is able to significantly
reduce the number of motif activity assignments that need to be evaluated. For a protein
with 15 motifs, the number of assignments evaluated is reduced from 215 = 32768 in
exhaustive search to 802 using our algorithm. Since majority of the computation is spent
on finding the activity assignments, this resulted in a 40 fold reduction in running time.
Predicting protein-protein interactions. We test our model by evaluating its performance
in predicting interactions. We test this performance using 5-fold cross validation on the set
of interacting and non-interacting protein pairs. In each fold, we train a model and predict
P (Tij .O) = true for pairs Pi , Pj in the held-out interactions.
Many motif pairs are over-represented in interacting proteins. We thus compare our
method to a baseline method that ranks pairs of proteins on the basis of the maximum
enrichment of over-represented motif pairs (see [18] for details). We also compare it to
a model where all motifs are set to be active; this is analogous to the method of Deng
et al. [8]. For completeness, we compare the two variants of the model using data on the
domain (Pfam and ProDom [19]) content of the proteins as well as the Prosite motif content.
Proportion of all interactions
predicted to interact
1
0.8
0.6
Association (Sprinzak & Margalit)
0.4
Prosite motif (Allow inactive motif. P.A = {0, 1})
Prosite motif (All motifs active. P.A = 1)
0.2
Pfam&ProDom (All domains active. P.A = 1.
Deng et al)
0
0
0.2
0.4
0.6
0.8
1
Proportion of all non-interactions predicted to interact
(a)
(b)
Figure 2: (a) ROC curve for different methods. The X-axis is the proportion of all non-interacting
protein pairs in the training data predicted to interact. Y-axis is the proportion of all interacting protein
pairs in the training data predicted to interact. Points are generated using different cutoff probabilties.
A larger area under the curve indicates better prediction. Our method (square marker) outperforms
all other methods. (b) Two protein chains that form a part of the 1ryp complex in PDB, interacting at
the site of two short sequence motifs.
The ROC curves in Fig. 2(a) show that our method outperforms the other methods, and that
the additional degree of freedom of allowing motifs to be inactive is essential. These results
validate our modeling assumptions; they also show that our method can be used to suggest
new interactions and to assign confidence levels on observed interactions, which is much
needed in view of the inaccuracies and large fraction of missing interactions in current
interaction databases.
Evaluating predicted active motifs. A key feature of our approach is its ability to detect
pairs of interacting motifs. We evaluate these predictions against the data from Protein Data
Bank (PDB) [20], which contains some solved structures of interacting proteins Fig. 2(b).
While the PDB data is scarce, it provides the ultimate evaluation of our predicted active
motifs. We extracted all structures from PDB that have at least two co-crystallized chains,
and whose chains are nearly identical to yeast proteins. From the residues that are in contact
between two chains (distance < 5 Angstr?oms), we infer which protein motifs participate in
interactions. Among our training data, 105 proteins have co-crystallized structure in PDB.
On these proteins, our data contained a total of 620 motif occurrences, of which 386 are
predicted to be active. Among those motifs predicted to be active, 257 of them (66.6%)
are interacting in PDB. Among the 234 motifs predicted to be inactive, only 120 of them
(51.3%) are interacting. The chi-square p-value is 10?4 . On the residue level, our predicted
active motifs consist of 3736 amino acids, and 1388 of them (37.2%) are interacting. In
comparison, our predicted inactive motifs consist of 3506 amino acids, and only 588 of
them (16.0%) are interacting. This significant enrichment provides support for the ability
of our method to detect motifs that participate in interactions. In fact, the set of interactions
in PDB is only a subset of the interactions those proteins participate in. Therefore, the
actual rate of false positive active motifs is likely to be lower than we report here.
5
Discussion and Conclusions
In this paper, we presented a probabilistic model which explicitly encodes elements in the
protein sequence that mediate protein-protein interactions. By using a variant of the EM al-
gorithm and a branch-and-bound algorithm for the E-step, we make the learning procedure
tractable. Our result shows that our method successfully uncovers motif activities and binding affinities, and uses them to predict both protein interactions and specific binding sites.
The ability of our model to predict structural elements, without a full structure analysis,
provides support for the viability of our approach.
Our use of a probabilistic model provides us with a general framework to incorporate
different types of data into our model, allowing it to be extended in varies ways. First,
we can incorporate additional signals for protein interactions, such as gene expression data
(as in [9]), cellular location, or even annotations from the literature (as in [7]). We can
also integrate protein interaction data across multiple species; for example, we might try to
use the yeast interaction data to provide more accurate predictions for the protein-protein
interactions in fly [10].
References
[1] P. Uetz, et al. A comprehensive analysis of protein-protein interactions in saccharomyces cerevisiae. Nature, 403(6770):623?7, 2000. 0028-0836 Journal Article.
[2] H. W. Mewes, et al. Mips: a database for genomes and protein sequences. Nucleic Acids Res,
2002.
[3] I. Xenarios, et al.Dip ; the database of interacting proteins: a research tool for studying cellular
networks of protein interactions. Nucleic Acids Research, 30(1):303?305, 2002. (c) 2002 Inst.
For Sci. Info.
[4] P. Chakrabarti and J. Janin. Dissecting protein protein recognition sites. PROTEINS: Structure,
Function, and Genetics, 47:334?343, 2002.
[5] J. J. Gray, et al.Protein protein docking with simultaneous optimization of rigid-body displacement and side-chain conformations. Journal of Molecular Biology, 331:281?299, 2003.
[6] Y. Ofran and B. Rost. Predicted protein-protein interaction sites from local sequence information. FEBS Lett., 544(1-3):236?239, 2003.
[7] R. Jansen, et al. A bayesian networks approach for predicting protein-protein interactions from
genomic data. Science, 302:449?53, 2003.
[8] M. Deng, S. Mehta, F. Sun, and T. Chen. Inferring domain-domain interactions from proteinprotein interactions. Genome Res, 12(10):1540?8, 2002. 22253763 1088-9051 Journal Article.
[9] E. Segal, H. Wang, and D. Koller. Discovering molecular pathways from protein interaction and
gene expression data. Bioinformatics, 19 Suppl 1:I264?I272, 2003. 1367-4803 Journal Article.
[10] L. Giot, et al. A protein interaction map of drosophila melanogaster. Science, 302(5651):1727?
36, 2003.
[11] L. Falquet, et al. The PROSITE database, its status in 2002. Nucliec Acids Research, 30:235?
238, 2002.
[12] D. R. Caffrey, et al. Are protein protein interfaces more conserved in sequence than the rest of
the protein surface? Protein Science, 13:190?202, 2003.
[13] J. Pearl. Probabilistic Reasoning in Intelligent Systems. Morgan Kaufmann, 1988.
[14] D. Koller and A. Pfeffer. Probabilistic frame-based systems. In Proc. AAAI, pages 580?587,
1998.
[15] A.P. Dempster, N.M. Laird, and D.B. Rubin. Maximum likelihood from incomplete data via
the em algorithm. J. Roy. Stat. Soc., B(39):1?39, 1977.
[16] J. S. Yedidia, W. T. Freeman, and Y. Weiss. Generalized belief propagation. In NIPS, pages
689?695, 2000.
[17] M. Henrion. Search-based methods to bound diagnostic probabilities in very large belief nets.
In Uncertainty in Artificial Intelligence, pages 142?150, 1991.
[18] E. Sprinzak and H. Margalit. Correlated sequence-signatures as markers of protein-protein
interaction. Journal of Molecular Biology, 311:681?692, 2001.
[19] R. Apweiler, et al. The interpro database, an integrated documentation resource for protein
families, domains and functional sites. Nucleic Acids Res, 29(1):37?40, 2001. 1362-4962
Journal Article.
[20] H.M. Berman, et al. The protein data bank. Nucleic Acids Research, 28:235?242, 2000.
| 2696 |@word proportion:4 mehta:1 uncovers:1 reduction:1 initial:1 contains:5 fragment:1 score:6 denoting:2 prefix:1 outperforms:2 existing:1 current:4 must:3 intelligence:1 discovering:2 short:2 completeness:1 provides:4 iterates:1 location:2 daphne:1 viable:1 chakrabarti:1 pathway:1 inside:1 introduce:1 expected:1 indeed:1 roughly:1 p1:5 chi:1 freeman:1 globally:2 actual:3 considering:1 begin:1 discover:1 underlying:1 notation:1 maximizes:1 finding:6 guarantee:1 every:1 exactly:1 wrong:1 omit:2 appear:2 positive:3 bind:6 local:4 tends:1 switching:1 despite:1 encoding:1 path:1 black:1 might:3 initialization:1 examined:1 specifying:1 co:3 patterned:1 limited:1 procedure:5 displacement:1 area:1 drug:2 significantly:3 revealing:1 confidence:2 induce:1 pdb:7 protein:140 suggest:1 get:1 cannot:1 put:1 influence:1 map:6 deterministic:2 center:1 missing:2 focused:1 identifying:3 aab:10 importantly:2 fill:1 handle:1 analogous:1 us:2 element:3 roy:1 recognition:1 documentation:1 gorithm:1 database:7 pfeffer:1 observed:9 p5:5 fly:1 wang:2 solved:2 capture:1 worst:1 region:1 t12:1 connected:2 sun:1 decrease:2 removed:1 intuition:3 dempster:1 inaccessible:1 complexity:1 signature:1 depend:1 solving:1 abd:1 asa:2 basis:1 easily:3 joint:4 represented:2 train:2 describe:1 artificial:1 crystallized:3 exhaustive:1 whose:2 stanford:5 larger:1 otherwise:4 ability:5 statistic:1 noisy:3 laird:1 final:1 sequence:14 apweiler:1 net:1 took:1 propose:2 interaction:94 interconnected:1 product:1 p4:3 relevant:1 achieve:1 validate:2 convergence:1 empty:1 optimum:1 assessing:1 produce:1 ben:1 help:1 spent:1 completion:1 stat:1 ij:1 conformation:1 eq:1 strong:2 throw:1 coverage:1 c:2 predicted:13 implies:1 soc:1 berman:1 p2:6 saved:1 attribute:1 require:1 premise:1 assign:1 drosophila:1 repertoire:1 adb:1 predict:4 biochemistry:1 achieves:1 purpose:2 estimation:1 proc:1 applicable:1 currently:1 successfully:1 tool:1 reflects:1 genomic:1 always:1 cerevisiae:2 aim:1 pn:1 encode:2 focus:1 refining:1 saccharomyces:1 rank:1 likelihood:5 indicates:1 contrast:1 baseline:1 detect:2 inst:1 inference:4 motif:123 rigid:1 typically:1 entire:2 a0:6 initially:1 spurious:4 koller:4 hidden:5 margalit:2 buried:1 dissecting:1 overall:1 among:3 denoted:1 jansen:1 initialize:1 construct:1 washington:2 bab:13 biology:3 represents:2 identical:1 throughput:4 nearly:1 others:1 report:2 simplify:2 inherent:1 few:2 intelligent:1 randomly:1 densely:2 resulted:1 comprehensive:1 phase:2 attempt:1 ab:10 freedom:1 highly:2 possibility:1 evaluation:1 behind:2 held:1 chain:5 accurate:2 incomplete:1 divide:1 initialized:2 circle:2 re:3 uncertain:1 modeling:1 assignment:39 a6:2 loopy:1 cost:1 subset:1 examining:1 dependency:1 varies:1 prosite:6 combined:1 stay:1 probabilistic:10 physic:1 quickly:1 aaai:1 resort:1 leading:1 account:2 segal:1 explicitly:1 ad:3 depends:1 try:3 picked:1 view:1 analyze:1 red:1 start:1 annotation:1 square:2 acid:7 kaufmann:1 efficiently:1 correspond:1 identify:2 yellow:1 bayesian:3 explain:3 simultaneous:1 against:1 associated:3 dataset:2 hur:1 knowledge:1 appears:2 higher:2 specify:1 wei:1 evaluated:5 pfam:2 correlation:1 until:1 working:1 marker:2 propagation:2 defines:1 quality:1 gray:2 yeast:3 hypothesized:1 contain:2 true:24 hence:4 illustrated:1 numerator:2 during:1 t25:1 generalized:1 bdb:1 complete:1 interface:1 reasoning:1 recently:1 common:2 functional:1 physical:3 discussed:3 association:1 significant:1 surface:4 add:4 feb:1 posterior:3 manipulation:1 rockefeller:1 binary:1 continue:1 integrated:1 conserved:3 morgan:1 fortunately:1 somewhat:1 additional:5 greater:1 deng:4 determine:2 maximize:4 converge:3 signal:1 branch:6 full:3 multiple:1 infer:2 reduces:2 cross:1 molecular:3 prediction:3 involving:1 basic:2 variant:3 iteration:2 tailored:1 suppl:1 cell:1 residue:2 addressed:2 rest:4 induced:2 tend:1 structural:1 presence:2 superset:1 mips:2 hb:2 switch:1 affect:1 viability:1 reduce:2 idea:1 inactive:11 whether:5 expression:2 inadequacy:1 ultimate:1 algebraic:1 tij:30 iterating:1 reduced:1 diagnostic:1 per:1 key:4 douglas:1 pj:13 cutoff:1 graph:1 fraction:3 sum:1 uncertainty:1 place:3 family:1 reasonable:2 p3:2 decision:1 entirely:1 bound:7 guaranteed:1 fold:3 g:1 activity:12 occur:2 precisely:1 encodes:1 min:1 relatively:1 department:3 developing:1 poor:1 across:2 terminates:1 em:7 making:1 explained:3 intuitively:1 computationally:1 resource:1 turn:3 describing:1 mechanism:1 needed:1 tractable:1 subnetworks:1 studying:1 highthroughput:1 yedidia:1 apply:2 observe:2 away:1 appropriate:1 occurrence:13 rost:1 robustness:1 assumes:1 remaining:2 binomial:1 running:1 exploit:3 contact:1 objective:6 occurs:4 eran:2 costly:1 affinity:16 distance:1 sci:1 entity:3 majority:1 participate:4 cellular:4 reason:1 relationship:1 illustration:1 ratio:1 unfortunately:1 potentially:4 info:1 negative:1 rise:1 design:3 prodom:2 brutlag:2 allowing:2 observation:2 nucleic:4 relational:2 extended:1 janin:1 frame:1 interacting:17 enrichment:2 pair:42 required:2 specified:1 pearl:1 inaccuracy:1 nip:1 address:3 able:1 below:1 pattern:2 including:1 max:1 explanation:1 belief:3 reliable:1 event:1 difficulty:1 predicting:6 scarce:2 improve:2 imply:1 axis:2 carried:1 ryp:1 mediated:1 docking:1 prior:3 literature:1 determining:1 relative:3 fully:1 oms:1 validation:1 integrate:1 degree:1 sufficient:1 article:4 rubin:1 bank:2 pi:42 genetics:1 repeat:1 side:1 allow:3 understand:3 aad:1 wide:3 dip:2 curve:3 lett:1 evaluating:2 genome:6 computes:1 doesn:1 correlate:1 melanogaster:1 approximate:6 pruning:1 status:1 proteinprotein:1 keep:1 gene:2 active:40 instantiation:2 conclude:1 search:6 decomposes:1 learn:1 terminate:1 nature:1 ca:2 interact:7 complex:2 domain:7 main:1 noise:1 arise:2 mediate:2 amino:2 body:1 site:19 fig:6 roc:2 ny:1 inferring:1 wish:2 exponential:2 candidate:2 specific:4 showing:1 evidence:1 dominates:1 exists:2 intractable:2 essential:1 false:16 adding:3 consist:2 chen:1 simply:1 appearance:1 likely:3 contained:1 binding:16 aa:14 extracted:1 conditional:1 goal:3 presentation:1 twofold:1 content:2 change:1 henrion:2 typical:1 except:3 determined:1 specifically:1 total:2 called:1 specie:1 experimental:1 t15:1 indicating:1 support:2 arises:1 relevance:1 bioinformatics:1 probabilties:1 incorporate:2 evaluate:2 correlated:2 |
1,868 | 2,697 | Theory of Localized Synfire Chain:
Characteristic Propagation Speed of
Stable Spike Patterns
Kosuke Hamaguchi
RIKEN Brain Science Institute
Wako, Saitama 351-0198, JAPAN
[email protected]
Masato Okada
Dept. of Complexity Science and
Engineering, University of Tokyo,
Kashiwa, Chiba, 277-8561, JAPAN
[email protected]
Kazuyuki Aihara
Institute of Industrial Science, University of Tokyo &
ERATO Aihara Complexity Modeling Project JST
Meguro, Tokyo 153-8505, JAPAN
[email protected]
Abstract
Repeated spike patterns have often been taken as evidence for the synfire
chain, a phenomenon that a stable spike synchrony propagates through
a feedforward network. Inter-spike intervals which represent a repeated
spike pattern are influenced by the propagation speed of a spike packet.
However, the relation between the propagation speed and network structure is not well understood. While it is apparent that the propagation
speed depends on the excitatory synapse strength, it might also be related
to spike patterns. We analyze a feedforward network with Mexican-Hattype connectivity (FMH) using the Fokker-Planck equation. We show
that both a uniform and a localized spike packet are stable in the FMH
in a certain parameter region. We also demonstrate that the propagation
speed depends on the distinct firing patterns in the same network.
1 Introduction
Neurons transmit information through spikes, but how the information is encoded remains a
matter of debate. The classical view is that the firing rates of neurons encode information,
while a recent view is that spatio-temporal spike patterns encode the information. For
example, the synchrony of spikes observed in the cortex is thought to play functional roles
in cognitive functions [1]. The mechanism of synchrony has been studied theoretically
for several neural network models. Especially, the model of spike synchrony propagation
through a feedforward network is called the synfire chain [2].
The mechanism of generating synchrony in a feedforward network can be described as follows. When feedforward connections are homogeneous with excitatory efficacy as a whole,
?
firing
rate
synaptic
input
r(?? , t)
?in (?, t)
?
....
??
position
?
inhibitory excitatory
- +
W(? - ??)
Mexican-Hat-type Connectivity
Figure 1: Network architecture. Each layer consists of N neuron arranged in a circle. Each
neuron projects its axon to a post-synaptic layer with Mexican-Hat-type connectivity.
post-synaptic neurons accept similar synaptic inputs. If neurons receive similar temporally
modulated inputs, the resultant spike timings will be also similar, or roughly synchronized
even though the membrane potentials fluctuate because of noise [3]. The question in a feedforward network is, whether the timing of spikes within a layer becomes more synchronized
or not as the spike packet propagates through a sequence of neural layers. Detailed analyses
of the activity propagation in feedforward networks have shown that homogeneous feedforward networks with excitatory synapses have a stable spike synchrony propagation mode
[4, 5, 6]. Neurons, however, are embedded in more structured networks with excitatory and
inhibitory synapses. Thus, such network structure would generate inhomogeneous inputs
to neurons and whether spike synchrony is stable is not a trivial issue.
One simple way to detect the synfire chain phenomena would be to record from several
neurons and find significant repeated patterns. If a spike packet propagates through a feedforward network, a statistically significant number of spike pairs would be found that have
fixed relative time lags (or inter-spike intervals (ISIs).) Such correlated activity has been
experimentally observed in the anterior forebrain pathway of songbirds [7], in the prefrontal cortex of primates [8], both in vivo and in vitro [9], and in an artificially constructed
network in vitro [10]. To generate fixed ISIs by spike packet propagation, the propagation speed of a spike packet must be constant over several trials. The speed depends on
spike patterns as well as the structure of the network. Conventional homogeneous feedforward networks have only one stable spike pattern, namely a spatially uniform synchronized
activity, but structured networks can generally produce spatially inhomogeneous spike patterns. In those networks, the relation between propagation speed and differences in the
spike pattern is not well understood.
It is therefore an important problem to study a biologically realistic, structured network in
the context of the synfire chain. Among suggested network structures, Mexican-Hat-type
(MH) connectivity is one of the most widely accepted as being representative of connectivity in the cortex [11]. Studies of a feedforward network with the MH connectivity (FMH)
have been reported [12]. In a FMH, a localized activity propagates through the network, and
this network is preferable to a homogeneous feedforward network because it can transmit
analog information regarding position [12], and both position and intensity [13]. However,
no detailed analytical work on the structured synfire chain has been reported. In this paper,
we use the Fokker-Planck equation to analyze the FMH. The method of the Fokker-Planck
equation enables us to analyze the collective behavior of the membrane potentials in an
identical neural population [14, 15]. When it is applied to the synfire chain [5], the detailed
analysis of the flow diagram, the effect of the membrane potential distribution on the spike
packet evolution, and the interaction of spike packets is possible [5].
This paper thus examines the feedforward neural network model with Mexican-Hat-type
connectivity. Our strategy is, first, to describe the evolution of firing states through order
parameters, which allows us to measure the macroscopic quantity of the network. Second,
we relate the input order parameters to the output ones through the Fokker-Planck equation.
Finally, we analyze the evolution of spike packets with various shapes, and investigate
stable firing patterns and their propagation speeds.
2 Model
We analyze the dynamics of a structured feedforward network composed of identical single compartment, Leaky Integrate-and-Fire (LIF) neurons. Each neuron is aligned in a ring
neural layer, and projects its axon to the next neural layer with the Mexican-Hat-type connectivity (Fig.1). The input to one neuron generally includes both outputs from pre-synaptic
neurons and a random noisy synaptic current from ongoing activity. If we assume that the
thousands of synaptic background inputs that connect to one neuron are independent Poissonian, instantaneous synapse, and have small amplitude of post synaptic potential (PSP),
we can approximate the sum of noisy background inputs as a Gaussian white noise fluctuating around the mean of the total noisy inputs. The membrane potential of a neuron at
position ? at time t, which receives many random Poisson excitatory and inhibitory inputs,
can be approximately described through a stochastic differential equation as follows:
dv(?, t)
v(?, t)
?
C
= Iin
+?
? + D0 ?(t),
(1)
(?, t) ?
dt
R
Z ?
d?0
?
Iin
W (? ? ?0 )r? (?0 , t),
(2)
(?, t) =
?? 2?
Z 0
r? (?, t) =
dt0 r(?, t ? t0 )?(t0 ),
(3)
??
W (?) = W0 + W1 cos(?),
(4)
?
where C is membrane capacitance, R is membrane resistance, I in (?, t) is the synaptic
current to a neuron at position ?, ?
? is proportional to the mean of the total noisy input, and
?(t) is a Gaussian random variable with h?(t)i = 0 and h?(t)?(t0 )i = ?(t ? t0 ). Here, D0 is
?
the amplitude of Gaussian noise. The input current Iin
(?, t) is obtained from the weighted
?
sum of output currents r (?, t) generated by pre-synaptic neurons. The synaptic current is
derived from the convolution of its firing rate r(?, t) with the PSP time course ?(t). Here,
?(t) = ??2 t exp(??t) where ? is chosen such that a single EPSP generates a 0.0014
mV elevation from the resting potential. The Mexican-Hat-type connectivity consists of
a uniform term W0 and a spatial modulation term W1 cos(?). We set the reset potential
and threshold potential as V0 and Vth , respectively. We start simulations from stationary
distribution of membrane potentials. The input to the initial layer is formulated in terms of
the firing rate on the virtual pre-synaptic layer,
(t ? t?)2
r0 + r1 cos(?)
?
exp(?
),
(5)
r(?, t) =
2? 2
2?? 2
where r0 and r1 are input parameters, and the temporal profile of activity is assumed to
be the Gaussian with ? = 1 and t? = 10. Throughout this paper, parameter values are
given as follows: C = 100 pF, R = 100 M?, Vth = 15 mV, D0 = 100, ?
? = 0.075 pA,
V0 = 0 mV, Vth = 15 mV, ? = 2, and ? = 0.00017. Space is divided into 50 regions
for the Fokker-Planck equation approach, and 10000 LIF neurons per layer are used in
simulations.
C
A
0.2 averaged membrane
potential distribution
mV
10
0
?10
P? = 0 (v,t)
10
0
?10
5
LIF
10
time [ms]
15
20
[P(v,t)]?
B
0
0.1
t=10.5 [ms]
mV
PDE
P?(v,t=10.5)
??
?
0
?
0
?15 ?10
V0
Vth
0
mV
10
15
Figure 2: A: Dynamics of membrane potential distribution at ? = 0. B: A snapshot of
the membrane potential distributions for a range of [?? ?] at t = 10.5. C: A snapshot
of the membrane potential distribution averaged with position ?. Results by a numerical
simulation (LIF) and the Fokker-Planck equation (PDE) are shown.
3 Theory
The prerequisites for a full description of the network activity are time series of order
parameters at an arbitrary time t defined as follows:
Z
p
1
r0 (t) =
(6)
d? r(?, t),
r1 (t) = (rc (t))2 + (rs (t))2 ,
2?
Z
Z
1
1
rc (t) =
d? cos(?)r(?, t),
rs (t) =
d? sin(?)r(?, t),
(7)
2?
2?
where r0 (t) is the population firing rate of the neuron population, and r c (t) and rs (t) are
coefficients of the Fourier transformation of the spatial firing pattern which represent the
spatial eccentricity of activity at time t. rc (t) and rs (t) depends on the position of the
localized activity, but r1 (t) does not. These order parameters play an important role in two
ways. First, we can describe the input to the next layer neuron at ? in a closed form of order
parameters. Second, their time integrals, which will be introduced later, become indices of
the spike packet shape. Input currents are described with the order parameters as follows:
?
Iin
(?, t) = W0 r0? (t) + W1 (rc? (t) cos(?) + rs? (t) sin(?)) ,
Z 0
?
r{0,c,s}
(t) =
dt0 ?(t0 ) r{0,c,s} (t ? t0 ).
(8)
(9)
??
Given the time sequence of order parameters in the pre-synaptic layer, the order parameters
in the post-synaptic layer are obtained through the following calculations. The analytical
method we use here is the Fokker-Planck equation which describes the distribution of the
membrane potential of a pool of identical neurons as the probability density P ? (v, t) of
voltage v at time t. The suffix ? denotes that this neuron population is located at position ?.
We assume that there are a large number of neurons at position ?. Equation (1) is equivalent
to the Fokker-Planck equation [16] within the limit of a large neuron number N ? ?,
?
?
P? (v, t) =
J? (v, t) + ?(v ? V0 )J? (Vth , t),
?t
?v
v
I ? (?, t) + ?
?
?
J? (v, t) =
? in
+
D P? (v, t),
?
C
?v
(10)
(11)
where J? (v, t) is a probability flux and D =
1
2
D0
C
2
. Boundary conditions are
P? (Vth , t) = 0,
r(?, t) =
J? (V0+ , t)
(12)
?
J? (V0? , t)
= J? (Vth , t).
(13)
Equation (12) is the absorbing boundary condition at the threshold potential, and Eq. (13) is
the current source at the reset potential. From Eq. (13), we obtain the firing rate of a postsynaptic neuron r(?, t). The Fokker-Planck equations are solved based on the modified
Chang-Cooper algorithm [17].
Figure 2 shows the actual distribution of the initial layer?s membrane potentials and their
dynamics which accepts virtual layer activity with parameter r0 = 500 and r1 = 350.
Figure 2A shows the evolution of the probability density P?=0 (v, t). From white to black,
the probability becomes higher. Figure 2B is a snapshot of the probability density at time
t = 10.5 over the region from ?? to ?. As a result of a localized input injection, part
of neuronal membrane potentials is strongly depolarized. The membrane potential distribution averaged over the neural layer is illustrated in Fig. 2C. It shows the consistency
between the numerical simulations and the Fokker-Planck equations. The probability flow
dropping out from the threshold potential Vth is a firing rate. Combined with these firing
rates at each position ? and definitions of order parameters in Eqs. (6)-(7), the order parameters on the post-synaptic neural layer are again calculated. The closed forms of order
parameters have been obtained.
Spatio-temporal patterns of firing rates and dynamics of order parameters in response to
a localized input (r0 = 600, r1 = 300) and a uniform input (r0 = 900, r1 = 0) are
shown in Fig. 3. When an input is spatially localized, the spatio-temporal firing pattern
is localized with a slightly distorted shape (Fig. 3A). On the other hand, when a uniform
input is applied, the spatio-temporal firing pattern is uniform as illustrated in Fig. 3B. We
show an example of the time course of r0 (t) and r1 (t) in Fig. 3C and 3D for both the
numerical simulation of 10, 000 LIF neurons and the Fokker-Planck equation. Elevation of
time course of r1 (t) in Fig. 3C indicates the localized firing. In contrast, the uniform input
generates no response in r1 (t) parameter as illustrated in Fig. 3D.
To quantitatively evaluate the spike packet shape and propagation speed, we define indices
r0 , r1 , and ?. r0 and r1 can be directly defined as
Z
Z
r0 = dt r0 (t) ? spontaneous firing rate, r1 = dt r1 (t).
(14)
r0 corresponds to the total population activity, and r1 corresponds to spatial eccentricity of
the activity. r0 and r1 are a natural extension of an index used in the study of the synfire
chain [4] in the sense that an index corresponds to the area of a time varying parameters
of the system, such as the population firing rate (r0 (t)) above the spontaneous firing rate,
or spatial eccentricity (r1 (t)). The basic idea of characterizing the spike packet was to
approximate the firing rate curve through a Gaussian function [4] as in Eq. (5). Here, the
approximated Gaussian curve is obtained by minimizing the mean squared error with r 0 (t)
and the Gaussian. We also use the index ? obtained from the variance of the Gaussian, and
t? as an index for the arrival time of the spike synchrony taken from the peak time of the
Gaussian (Fig. 3C).
4 Results
Our observation of the activities of the FMH with various parameter sets reveals two types
of stable spike packets. Figure 4 shows the activity of the FMH with four characteristic
parameter sets of W0 and W1 . Here we use r0 = 500 and r1 = 350 for the upper figures
A
0
??
5
10
time [ms]
15
0
??
20
C
5
10
time [ms]
15
20
D
?
500
t
0
5
1000
FP
LIF
r0(t)
[spk/s]
1000 Gaussian
Approximation
[spk/s]
firing rate
?
?
?
B
firing rate
?
r1(t)
10
time [ms]
15
20
r0(t)
FP
LIF
500
r1(t)
0
5
10
time [ms]
15
20
Figure 3: Activity profiles in response to a localized input (A,C) and a uniform input (B,D).
A,B: Spatio-temporal pattern of the firing rates from neurons at position ?? to ?. C,D:
Time courses of order parameters (r0 (t), r1 (t)) calculated from numerical simulations of
a population of LIF neurons (squares and crosses) and the Fokker-Planck equation (solid
lines). The time course of order parameters in response to a single stimulus is approximated
by using a Gaussian function. In C, Gaussian approximation of r 0 (t) is also shown. The
variance of the Gaussian ? and the mean value t? are used as the indices of a spike packet.
as a localized input and r0 = 900 and r1 = 0 for the lower ones as a uniform input. The
common parameter is ? = 1 and t? = 2. When both W0 and W1 are small, no spike packet
can propagate (Non-firing). When the uniform activation term W 0 is sufficiently strong,
a uniform spike packet is stable (Uniform Activity). Note that even though the localized
input elicits localized spike packets with several layers, it finally decays to the uniform
spike packet. When the Mexican-Hat term W1 is strong enough, only the localized spike
packet is stable (Localized Activity). When W0 and W1 are balanced within a certain ratio,
there exists a novel firing mode where both the uniform and the localized spike packet are
stable depending on the initial layer input (Multi-stable). The results show that there are
four types of phase and two types of spike packet in the FMH. The stability of a spike packet
depends on W0 and W1 . In addition, the difference of the arrival times of propagating spike
packets in the 8th layer shown in the Multi-stable phase indicates that the propagation speed
of spike packets might differ.
It is apparent that the propagation speed depends on the strength of the excitatory synapse
efficacy, however, our results in the Multi-stable phase in Fig. 4 suggest that a spike pattern
also determines the propagation speed. To investigate this effect, we plotted propagation
time ?t?, the difference between propagation time t?post and t?pre for various W1 (Fig. 5B).
The speed is analyzed after the spike packet indices r0 , r1 and ? have converged. The
convergences of spike packets are shown in the flow diagram in Fig. 5A for (W 0 , W1 ) =
(1, 1.5) case. In Fig. 5B, each triangle indicates the speed of the localized activity, and each
circle corresponds to that of the uniform activity. Within the plotted region (W 1 = 1.4 ?
2), both the uniform and localized activities are stable, and no bursting activity is observed.
This indicates that as W1 rises the propagation speed of localized activity becomes higher.
In contrast, the propagation speed of the uniform activity does not depends on W 1 because
Layer
Uniform Input Localized Input
Non-firing
(W0,W1) = (0.7, 1)
1
2
3
4
5
6
7
8
1
2
3
4
5
6
7
8
0
5
10
15
Uniform Activity
(W0,W1) = (1, 0.6)
20 0
5
10
15
Localized Activity Multi-stable phase
(W0,W1) = (0.7, 2.5) (W0,W1) = (1, 1.4)
20 0
5
10
15
20 0
5
10
15
20
time [ms]
Figure 4: Four characteristic FMH with different stable states. The evolutions of firing rate
propagation are shown. The upper row panels show the response to an identical localized
pulse input, and the lower row panels show the response to a uniform pulse input.
uniform activity generates rc (t) = rs (t) = 0.
5 Summary
We have found that there are four phases in the W0 ? W1 parameter space; Non-firing,
Localized activity, Uniform activity, and Multi-stable phase. Multi-stable phase is the most
intriguing in that an identical network has completely different firing modes in response to
different initial inputs. In this phase, the effect of spike pattern on the propagation speed
of the spike packet can be directly studied. By the analysis of the Fokker-Planck equation,
we found that the propagation speed depends on the distinct firing patterns in the same
network. It implies that observation of repeated spike patterns requires an appropriately
controlled input if the network structure produces a multi-stable state. The characteristic
speed of the spike packet also suggests that the speed of information processing in the brain
depends on the spiking pattern, or the representation of the information.
Acknowledgment
This study is partially supported by the Advanced and Innovational Research Program in
Life Sciences, a Grant-in-Aid No. 15016023 for Scientific Research on Priority Areas
(C) Advanced Brain Science Project, a Grand-in-Aid No. 14084212, from the Ministry of
Education, Culture, Sports, Science, and Technology, the Japanese Government.
References
[1] C. M. Gray, P. K?onig, A. K. Engel, and W. Singer, ?Oscillatory responses in cat visual cortex exhibit inter-columnar synchronization which reflects global stimulus properties,? Nature,
vol. 338, pp. 334?337, 1989.
[2] M. Abeles, Corticonics: neural circuits of the cerebral cortex. Cambridge UP, 1991.
[3] Z. Mainen and T. Sejnowski, ?Reliability of spike timing in neocortical neurons,? Science,
vol. 268, pp. 1503?1506, 1995.
[4] M. Diesmann, M.-O. Gewaltig, and A. Aertsen, ?Stable propagation of synchronous spiking in
cortical neural networks,? Nature, vol. 402, pp. 529?533, 1999.
A
w0 = 1
B
1.2
local
uniform
-
? t [ms]
500
300
r1
Localized
100
Uniform
0
300
200
100
?/?
0
200
600
r0
1000
1
0.8
1.4
1.6
1.8
w1
2
Figure 5: A: Flow diagram with parameter (W0 , W1 ) = (1, 1.5) where both localized and
uniform spike packets are stable. Two attractors in high r1 region (Localized) and high r0
with r1 = 0 region (Uniform) are shown. A sequence of arrows indicates the evolution of
a spike packet in the (r0 , ?/?, r1 ) space. B: Plot of propagation time ?t? that is necessary
time for a spike packet to propagate one neural layer. These results indicate that localized
spike packets propagate slower than the uniform ones.
[5] H. C?ateau and T. Fukai, ?Fokker-planck approach to the pulse packet propagation in synfire
chain,? Neural Networks, vol. 14, pp. 675?685, 2001.
[6] W. M. Kistler and W. Gerstner, ?Stable propagation of activity pulses in populations of spiking
neurons,? Neural Comp., vol. 14, pp. 987?997, 2002.
[7] R. Kimpo, F. Theunissen, and A. Doupe, ?Propagation of correlated activity through multiple
stages of a neural circuit,? J. Neurosci., vol. 23, no. 13, pp. 5750?5761, 2003.
[8] M. Abeles, H. Bergman, E. Margalit, and E. Vaadia, ?Spatiotemporal firing patterns in the
frontal cortex of behaving monkeys,? J. Neurophysiol., vol. 70, pp. 1629?1638, 1993.
[9] Y. Ikegaya, G. Aaron, R. Cossart, D. Aronov, I. Lampl, D. Ferster, and R. Yuste, ?Synfire
chains and cortical songs: Temporal modules of cortical activity,? Science, vol. 304, pp. 559?
564, 2004.
[10] A. Reyes, ?Synchrony-dependent propagation of firing rate in iteratively constructed networks
in vitro,? Nature Neuroscience, vol. 6, pp. 593 ? 599, 2003.
[11] D. H. Hubel and T. N. Wiesel, ?Receptive fields, binocular interaction and functional architecture in the cat?s visual cortex,? J. Physiol., vol. 160, pp. 106?154, 1962.
[12] M. C. W. van Rossum, G. G. Turrigiano, and S. B. Nelson, ?Fast propagation of firing rates
through layered networks of noisy neurons,? J. Neurosci., vol. 22, pp. 1956?1966, 2002.
[13] K. Hamaguchi and K. Aihara, ?Quantitative information transfer through layers of spiking neurons connected by mexican-hat type connectivity,? Neurocomputing, vol. 58-60, pp. 85?90,
2004.
[14] D. J. Amit and N. Brunel, ?Dynamics of a recurrent network of spiking neurons before and
following learning,? Network: Computation in Neural Systems, vol. 8, no. 4, pp. 373?404,
1997.
[15] N. Brunel and V. Hakim, ?Fast Global Oscillations in Networks of Integrate-and-Fire Neurons
with Low Firing Rates,? Neural Comp., vol. 11, no. 7, pp. 1621?1671, 1999.
[16] H. Risken, The Fokker-Planck Equation. Springer-Verlag, 1996.
[17] J. S. Chang and G. Cooper, ?A practical difference scheme for fokker-planck equations,? J.
Comp. Phys., vol. 6, pp. 1?16, 1970.
| 2697 |@word trial:1 wiesel:1 r:6 propagate:3 simulation:6 pulse:4 solid:1 initial:4 series:1 efficacy:2 mainen:1 wako:1 current:7 anterior:1 activation:1 intriguing:1 must:1 physiol:1 numerical:4 realistic:1 shape:4 enables:1 plot:1 stationary:1 record:1 rc:5 constructed:2 differential:1 become:1 consists:2 pathway:1 theoretically:1 inter:3 roughly:1 isi:2 behavior:1 multi:7 brain:5 actual:1 pf:1 becomes:3 project:4 panel:2 circuit:2 monkey:1 transformation:1 temporal:7 quantitative:1 preferable:1 onig:1 grant:1 planck:16 rossum:1 before:1 engineering:1 understood:2 timing:3 local:1 limit:1 firing:34 modulation:1 approximately:1 might:2 black:1 studied:2 bursting:1 suggests:1 co:5 range:1 statistically:1 averaged:3 acknowledgment:1 practical:1 area:2 thought:1 pre:5 suggest:1 layered:1 context:1 conventional:1 equivalent:1 examines:1 stability:1 population:8 transmit:2 spontaneous:2 play:2 homogeneous:4 bergman:1 pa:1 approximated:2 located:1 theunissen:1 observed:3 role:2 module:1 solved:1 thousand:1 region:6 connected:1 balanced:1 complexity:2 dynamic:5 completely:1 triangle:1 spk:2 neurophysiol:1 mh:2 various:3 cat:2 riken:3 distinct:2 fast:2 describe:2 sejnowski:1 apparent:2 encoded:1 lag:1 widely:1 dt0:2 noisy:5 sequence:3 vaadia:1 turrigiano:1 analytical:2 interaction:2 epsp:1 reset:2 aligned:1 meguro:1 description:1 convergence:1 eccentricity:3 r1:27 produce:2 generating:1 ring:1 depending:1 recurrent:1 ac:1 propagating:1 eq:4 strong:2 implies:1 indicate:1 synchronized:3 differ:1 inhomogeneous:2 tokyo:4 hammer:1 stochastic:1 packet:33 jst:1 virtual:2 kistler:1 education:1 government:1 elevation:2 extension:1 around:1 sufficiently:1 exp:2 engel:1 weighted:1 reflects:1 gaussian:13 cossart:1 modified:1 fluctuate:1 varying:1 voltage:1 encode:2 derived:1 indicates:5 industrial:1 contrast:2 detect:1 sense:1 dependent:1 suffix:1 accept:1 margalit:1 relation:2 issue:1 among:1 spatial:5 lif:8 field:1 corticonics:1 identical:5 stimulus:2 quantitatively:1 composed:1 neurocomputing:1 phase:8 fire:2 attractor:1 aronov:1 investigate:2 analyzed:1 chain:10 integral:1 necessary:1 culture:1 circle:2 plotted:2 modeling:1 saitama:1 uniform:27 reported:2 connect:1 spatiotemporal:1 abele:2 combined:1 density:3 peak:1 grand:1 pool:1 connectivity:10 w1:18 again:1 squared:1 prefrontal:1 priority:1 cognitive:1 japan:3 potential:20 includes:1 coefficient:1 matter:1 mv:7 depends:9 later:1 view:2 closed:2 analyze:5 start:1 synchrony:9 vivo:1 square:1 compartment:1 variance:2 characteristic:4 comp:3 converged:1 oscillatory:1 synapsis:2 influenced:1 phys:1 synaptic:15 definition:1 pp:15 resultant:1 amplitude:2 higher:2 dt:3 response:8 synapse:3 arranged:1 though:2 strongly:1 stage:1 binocular:1 hand:1 receives:1 synfire:10 propagation:30 mode:3 gray:1 scientific:1 effect:3 evolution:6 spatially:3 iteratively:1 illustrated:3 white:2 sin:2 erato:1 songbird:1 m:8 neocortical:1 demonstrate:1 reyes:1 iin:4 instantaneous:1 novel:1 common:1 absorbing:1 functional:2 spiking:5 vitro:3 jp:3 cerebral:1 analog:1 resting:1 significant:2 cambridge:1 consistency:1 reliability:1 stable:23 cortex:7 behaving:1 v0:6 recent:1 certain:2 verlag:1 life:1 ministry:1 r0:25 full:1 multiple:1 d0:4 calculation:1 cross:1 pde:2 dept:1 divided:1 post:6 controlled:1 basic:1 poisson:1 represent:2 receive:1 background:2 addition:1 interval:2 diagram:3 source:1 fukai:1 macroscopic:1 appropriately:1 depolarized:1 flow:4 feedforward:14 enough:1 architecture:2 regarding:1 idea:1 masato:1 fmh:9 whether:2 t0:6 synchronous:1 song:1 resistance:1 generally:2 detailed:3 generate:2 inhibitory:3 kosuke:1 neuroscience:1 per:1 dropping:1 vol:15 four:4 threshold:3 sum:2 distorted:1 throughout:1 oscillation:1 layer:22 risken:1 activity:31 strength:2 diesmann:1 generates:3 fourier:1 speed:21 injection:1 structured:5 membrane:15 psp:2 describes:1 slightly:1 postsynaptic:1 primate:1 biologically:1 aihara:4 dv:1 taken:2 equation:18 remains:1 forebrain:1 mechanism:2 singer:1 prerequisite:1 fluctuating:1 slower:1 hat:8 lampl:1 denotes:1 especially:1 amit:1 classical:1 capacitance:1 question:1 quantity:1 spike:57 strategy:1 receptive:1 aertsen:1 exhibit:1 elicits:1 w0:14 nelson:1 trivial:1 index:8 ratio:1 minimizing:1 relate:1 debate:1 rise:1 collective:1 upper:2 neuron:34 convolution:1 snapshot:3 observation:2 arbitrary:1 intensity:1 introduced:1 pair:1 namely:1 connection:1 accepts:1 poissonian:1 suggested:1 pattern:23 fp:2 program:1 natural:1 advanced:2 scheme:1 technology:1 temporally:1 gewaltig:1 kazuyuki:1 relative:1 embedded:1 synchronization:1 yuste:1 proportional:1 localized:27 integrate:2 propagates:4 row:2 excitatory:7 course:5 summary:1 supported:1 hakim:1 institute:2 characterizing:1 leaky:1 van:1 chiba:1 boundary:2 calculated:2 curve:2 cortical:3 flux:1 approximate:2 global:2 hubel:1 reveals:1 sat:1 assumed:1 spatio:5 nature:3 okada:2 transfer:1 gerstner:1 japanese:1 artificially:1 neurosci:2 arrow:1 whole:1 noise:3 profile:2 arrival:2 repeated:4 neuronal:1 fig:13 representative:1 cooper:2 axon:2 aid:2 position:11 ikegaya:1 decay:1 evidence:1 exists:1 columnar:1 kashiwa:1 vth:8 visual:2 partially:1 sport:1 chang:2 brunel:2 springer:1 fokker:16 corresponds:4 determines:1 formulated:1 ferster:1 experimentally:1 mexican:9 called:1 total:3 accepted:1 aaron:1 doupe:1 modulated:1 frontal:1 ongoing:1 evaluate:1 phenomenon:2 correlated:2 |
1,869 | 2,698 | Sharing Clusters Among Related Groups:
Hierarchical Dirichlet Processes
Yee Whye Teh(1) , Michael I. Jordan(1,2), Matthew J. Beal(3) and David M. Blei(1)
(1)
(3)
Computer Science Div., (2) Dept. of Statistics
Dept. of Computer Science
University of California at Berkeley
University of Toronto
Berkeley CA 94720, USA
Toronto M5S 3G4, Canada
{ywteh,jordan,blei}@cs.berkeley.edu
[email protected]
Abstract
We propose the hierarchical Dirichlet process (HDP), a nonparametric
Bayesian model for clustering problems involving multiple groups of
data. Each group of data is modeled with a mixture, with the number of
components being open-ended and inferred automatically by the model.
Further, components can be shared across groups, allowing dependencies
across groups to be modeled effectively as well as conferring generalization to new groups. Such grouped clustering problems occur often in
practice, e.g. in the problem of topic discovery in document corpora. We
report experimental results on three text corpora showing the effective
and superior performance of the HDP over previous models.
1 Introduction
One of the most significant conceptual and practical tools in the Bayesian paradigm is
the notion of a hierarchical model. Building on the notion that a parameter is a random
variable, hierarchical models have applications to a variety of forms of grouped or relational
data and to general problems involving ?multi-task learning? or ?learning to learn.? A
simple and classical example is the Gaussian means problem, in which a grand mean ?0
is drawn from some distribution, a set of K means are then drawn independently from a
Gaussian with mean ?0 , and data are subsequently drawn independently from K Gaussian
distributions with these means. The posterior distribution based on these data couples the
means, such that posterior estimates of the means are shrunk towards each other. The
estimates ?share statistical strength,? a notion that can be made precise within both the
Bayesian and the frequentist paradigms.
Here we consider the application of hierarchical Bayesian ideas to a problem in ?multi-task
learning? in which the ?tasks? are clustering problems, and our goal is to share clusters
among multiple, related clustering problems. We are motivated by the task of discovering
topics in document corpora [1]. A topic (i.e., a cluster) is a distribution across words while
documents are viewed as distributions across topics. We want to discover topics that are
common across multiple documents in the same corpus, as well as across multiple corpora.
Our work is based on a tool from nonparametric Bayesian analysis known as the Dirichlet
process (DP) mixture model [2, 3]. Skirting technical definitions for now, ?nonparametric?
can be understood simply as implying that the number of clusters is open-ended. Indeed,
at each step of generating data points, a DP mixture model can either assign the data point
to a previously-generated cluster or can start a new cluster. The number of clusters is a
random variable whose mean grows at rate logarithmic in the number of data points.
Extending the DP mixture model framework to the setting of multiple related clustering
problems, we will be able to make the (realistic) assumption that we do not know the
number of clusters a priori in any of the problems, nor do we know how clusters should be
shared among the problems.
When generating a new cluster, a DP mixture model selects the parameters for the cluster
(e.g., in the case of Gaussian mixtures, the mean and covariance matrix) from a distribution
G0 ?the base distribution. So as to allow any possible parameter value, the distribution
G0 is often assumed to be a smooth distribution (i.e., non-atomic). Unfortunately, if we
now wish to extend DP mixtures to groups of clustering problems, the assumption that G0
is smooth conflicts with the goal of sharing clusters among groups. That is, even if each
group shares the same underlying base distribution G0 , the smoothness of G0 implies that
they will generate distinct cluster parameters (with probability one). We will show that this
problem can be resolved by taking a hierarchical Bayesian approach. We present a notion
of a hierarchical Dirichlet process (HDP) in which the base distribution G0 for a set of DPs
is itself a draw from a DP. This turns out to provide an elegant and simple solution to the
problem of sharing clusters among multiple clustering problems.
The paper is organized as follows. In Section 2, we provide the basic technical definition
of DPs and discuss related representations involving stick-breaking processes and Chinese
restaurant processes. Section 3 then introduces the HDP, motivated by the requirement
of a more powerful formalism for the grouped data setting. As for the DP, we present
analogous stick-breaking and Chinese restaurant representations for the HDP. We present
empirical results on a number of text corpora in Section 5, demonstrating various aspects of
the HDP including its nonparametric nature, hierarchical nature, and the ease with which
the framework can be applied to other realms such as hidden Markov models.
2 Dirichlet Processes
The Dirichlet process (DP) and the DP mixture model are mainstays of nonparametric
Bayesian statistics (see, e.g., [3]). They have also begun to be seen in applications in machine learning (e.g., [7, 8, 9]). In this section we give a brief overview with an eye towards
generalization to HDPs. We begin with the definition of DPs [4]. Let (?, B) be a measurable space, with G0 a probability measure on the space, and let ?0 be a positive real number.
A Dirichlet process is the distribution of a random probability measure G over (?, B) such
that, for any finite partition (A1 , . . . , Ar ) of ?, the random vector (G(A1 ), . . . , G(Ar )) is
distributed as a finite-dimensional Dirichlet distribution:
(G(A1 ), . . . , G(Ar )) ? Dir ?0 G0 (A1 ), . . . , ?0 G0 (Ar ) .
(1)
We write G ? DP(?0 , G0 ) if G is a random probability measure distributed according to
a DP. We call G0 the base measure of G, and ?0 the concentration parameter.
The DP can be used in the mixture model setting in the following way. Consider a set
of data, x = (x1 , . . . , xn ), assumed exchangeable. Given a draw G ? DP(?0 , G0 ),
independently draw n latent factors from G: ?i ? G. Then, for each i = 1, . . . , n,
draw xi ? F (?i ), for a distribution F . This setup is referred to as a DP mixture model.
If the factors ?i were all distinct, then this setup would yield an (uninteresting) mixture
model with n components. In fact, the DP exhibits an important clustering property, such
that the draws ?i are generally not distinct. Rather, the number of distinct values grows as
O(log n), and it is this that defines the random number of mixture components.
There are several perspectives on the DP that help to understand this clustering property.
In this paper we will refer to two: the Chinese restaurant process (CRP), and the stickbreaking process. The CRP is a distribution on partitions that directly captures the clustering of draws from a DP via a metaphor in which customers share tables in a Chinese restaurant [5]. As we will see in Section 4, the CRP refers to properties of the joint distribution
of the factors {?i }. The stick-breaking process, on the other hand, refers to properties of
G, and directly reveals its discrete nature [6]. For k = 1, 2 . . ., let:
Qk?1
?k ? G0
?k0 ? Beta(1, ?0 )
?k = ?k0 l=1 (1 ? ?k0 ).
(2)
P?
Then with probability one the random measure defined by G = k=1 ?k ??k is a sample
from DP(?0 , G0 ). The construction for ?1 , ?2 , . . . in (2) can be understood as taking a
stick of unit length, and repeatedly breaking off segments of length ?k . The stick-breaking
construction shows that DP mixture models can be viewed as mixture models with a countably infinite number of components. To see this, identify each ?k as the parameter of the
k th mixture component, with mixing proportion given by ?k .
3 Hierarchical Dirichlet Processes
We will introduce the hierarchical Dirichlet process (HDP) in this section. First we describe the general setting in which the HDP is most useful?that of grouped data. We
assume that we have J groups of data, each consisting of nj data points (xj1 , . . . , xjnj ).
We assume that the data points in each group are exchangeable, and are to be modeled with
a mixture model. While each mixture model has mixing proportions specific to the group,
we require that the different groups share the same set of mixture components. The idea is
that while different groups have different characteristics given by a different combination
of mixing proportions, using the same set of mixture components allows statistical strength
to be shared across groups, and allows generalization to new groups.
The HDP is a nonparametric prior which allows the mixture models to share components.
It is a distribution over a set of random probability measures over (?, B): one probability
measure Gj for each group j, and a global probability measure G0 . The global measure G0
is distributed as DP(?, H), with H the base measure and ? the concentration parameter,
while each Gj is conditionally independent given G0 , with distribution Gj ? DP(?0 , G0 ).
To complete the description of the HDP mixture model, we associate each xji with a factor
?ji , with distributions given by F (?ji ) and Gj respectively. The overall model is given in
Figure 1 left, with conditional distributions:
G0 | ?, H ? DP(?, H)
?ji | Gj ? Gj
Gj | ?, G0 ? DP(?0 , G0 )
xji | ?ji ? F (?ji ) .
(3)
(4)
The stick-breaking construction
(2) shows that a draw of G0 can be expressed as a weighted
P?
sum of point masses: G0 = k=1 ?k ??k . This fact that G0 is atomic plays an important
role in ensuring that mixture components are shared across different groups. Since G0 is
the base distribution for the individual Gj ?s, (2) again shows that the atoms of the individual
Gj are samples from G0 . In particular, since G0 places non-zero mass only on the atoms
? = (?k )?
k=1 , the atoms of Gj must also come from ?, hence we may write:
P
P
G0 = ?
Gj = ?
(5)
k=1 ?k ??k
k=1 ?jk ??k .
Identifying ?k as the parameters of the k th mixture component, we see that each submodel
corresponding to distinct groups share the same set of mixture components, but have differing mixing proportions, ? j = (?jk )?
k=1 .
Finally, it is useful to explicitly describe the relationships between the mixing proportions
? and (? j )Jj=1 . Details are provided in [10]. Note that the weights ?j are conditionally independent given ? since each Gj is independent given G0 . Applying (1) to finite partitions
H
?
?0
?0
G1
?1i
global
?11
?13
G0
?0
G2
x1i
x2i
n1
?23 ?32
G3
?18 ?11
?16
? ?15
?11 ?13 ?12 ?12
14
?17
n3
?22
?21 ?21
?25
?23
?24
?23 ?22 ?26
?24
?13
x3i
n2
?31
?22
?12 ?3
?24 ?21
?3i
?2i
?2
?1
?36
?31 ?35
?31
?32
?34
?33
?32
?28 ?27
group j=2
group j=1
group j=3
Figure 1: Left: graphical model of an example HDP mixture model with 3 groups. Corresponding
to each DP node we also plot a sample draw from the DP using the stick-breaking construction.
Right: an instantiation of the CRF representation for the 3 group HDP. Each of the 3 restaurants has
customers sitting around tables, and each table is served a dish (which corresponds to customers in
the Chinese restaurant for the global DP).
of ?, we get ? j ? DP(?0 , ?), where we interpret ? and ? j as probability measures over
the positive integers. Hence ? is simply the putative mixing proportion over the groups.
We may in fact obtain an explicit stick-breaking construction for the ? j ?s as well. Applying
(1) to partitions ({1, . . . , k ? 1}, {k}, {k + 1, . . .}) of positive integers, we have:
Pk
0
0 Qk?1
0
?jk
? Beta ?0 ?k , ?0 1 ? l=1 ?l
?jk = ?jk
(6)
l=1 (1 ? ?jl ) .
4 The Chinese Restaurant Franchise
We describe an alternative view of the HDP based directly upon the distribution a HDP induces on the samples ?ji , where we marginalize out G0 and Gj ?s. This view directly leads
to an efficient Gibbs sampler for HDP mixture models, which is detailed in the appendix.
Consider, for one group j, the distribution of ?j1 , . . . , ?jnj as we marginalize out Gj . Recall that since Gj ? DP(?0 , G0 ) we can describe this distribution by describing how to
generate ?j1 , . . . , ?jnj using the CRP. Imagine nj customers (each corresponds to a ?ji )
at a Chinese restaurant with an unbounded number of tables. The first customer sits at the
first table. A subsequent customer sits at an occupied table with probability proportional
to the number of customers already there, or at the next unoccupied table with probability
proportional to ?0 . Suppose customer i sat at table tji . The conditional distributions are:
X
P njt
tji | tj1 , . . . , tji?1 , ?0 ?
?t + P 0 n?00 +?0 ?tnew ,
(7)
0 n 0 +?0
t
t
jt
t
jt
where njt is the number of customers currently at table t. Once all customers have sat down
the seating plan corresponds to a partition of ?j1 , . . . , ?jnj . This is an exchangeable process in that the probability of a partition does not depend on the order in which customers
sit down. Now we associate with table t a draw ?jt from G0 , and assign ?ji = ?jtji .
Performing this process independently for each group j, we have now integrated out all the
Gj ?s, and have an assignment of each ?ji to a sample ?jtji from G0 , with the partition
structures given by CRPs. Notice now that all ?jt ?s are simply i.i.d. draws from G0 , which
is again distributed according to DP(?, H), so we may apply the same CRP partitioning
process to the ?jt ?s. Let the customer associated with ?jt sit at table kjt . We have:
X
P mk
kjt | k11 , . . . , k1n1 , k21 , . . . , kjt?1 , ? ?
?k + P 0 m? 0 +?0 ?knew . (8)
0 m 0 +?
k
k
jk
k
k
Perplexity on test abstacts of LDA and HDP mixture
Posterior over number of topics in HDP mixture
1050
15
LDA
HDP Mixture
Number of samples
Perplexity
1000
950
900
850
10
5
800
750
10
20
30
40
50 60 70 80 90 100 110 120
Number of LDA topics
0
61 62 63 64 65 66 67 68 69 70 71 72 73
Number of topics
Figure 2: Left: comparison of LDA and HDP mixture. Results are averaged over 10 runs, with error
bars being 1 standard error. Right: histogram of the number of topics the HDP mixture used over 100
posterior samples.
Finally we associate with table k a draw ?k from H and assign ?jt = ?kjt . This completes
the generative process for the ?ji ?s, where we marginalize out G0 and Gj ?s. We call this
generative process the Chinese restaurant franchise (CRF). The metaphor is as follows: we
have J restaurants, each with nj customers (?ji ?s), who sit at tables (?jt ?s). Now each table
is served a dish (?k ?s) from a menu common to all restaurants. The customers are sociable,
prefering large tables with many customers present, and also prefer popular dishes.
5 Experiments
We describe 3 experiments in this section to highlight the various aspects of the HDP: its
nonparametric nature; its hierarchical nature; and the ease with which we can apply the
framework to other models, specifically the HMM.
Nematode biology abstracts. To demonstrate the strength of the nonparametric approach
as exemplified by the HDP mixture, we compared it against latent Dirichlet allocation
(LDA), which is a parametric model similar in structure to the HDP [1]. In particular,
we applied both models to a corpus of nematode biology abstracts1 , evaluating the perplexity of both models on held out abstracts. Here abstracts correspond to groups, words
correspond to observations, and topics correspond to mixture components, and exchangeability correspond to the typical bag-of-words assumption. In order to study specifically the
nonparametric nature of the HDP, we used the same experimental setup for both models2 ,
except that in LDA we had to vary the number of topics used between 10 and 120, while
the HDP obtained posterior samples over this automatically.
The results are shown in Figure 2. LDA performs best using between 50 and 80 topics,
while the HDP performed just as well as these. Further, the posterior over the number of
topics used by HDP is consistent with this range. Notice however that the HDP infers the
number of topics automatically, while LDA requires some method of model selection.
NIPS sections. We applied HDP mixture models to a dataset of NIPS 1-12 papers organized into sections3 . To highlight the transfer of learning achievable with the HDP, we
1
Available at http://elegans.swmed.edu/wli/cgcbib. There are 5838 abstracts in total. After removing
standard stop words and words appearing less than 10 times, we are left with 476441 words in total
and a vocabulary size of 5699.
2
In both models, we used a symmetric Dirichlet distribution with weights of 0.5 for the prior H
over topic distributions, while the concentration parameters are integrated out using a vague gamma
prior. Gibbs sampling using the CRF is used, while the concentration parameters are sampled using
a method described in [10]. This also applies to the NIPS sections experiment on next page.
3
To ensure we are dealing with informative words in the documents, we culled stop words as well
Average perplexity over NIPS sections of 3 models
Generalization from LT, AA, AP to VS
6000
5000
M1: additional sction ignored
M2: flat, additional section
M3: hierarchical, additional section
5500
Perplexity
Perplexity
5000
LT
AA
AP
4500
4500
4000
4000
3500
3500
3000
3000
2500
0
10
20
30
40
50
60
Number of VS training documents
70
80
2500
0
10
20
30
40
50
60
Number of VS training documents
70
80
Figure 3: Left: perplexity of test VS documents given training documents from VS and another
section for 3 different models. Curves shown are averaged over the other sections and 5 runs. Right:
perplexity of test VS documents given LT, AA and AP documents respectively, using M3, averaged
over 5 runs. In both, the error bars are 1 standard error.
show improvements to the modeling of a section when the model is also given documents
from another section. Our test section is always the VS (vision sciences) section, while
the additional section is varied across the other eight. The training set always consist of
80 documents from the other section (so that larger sections like AA (algorithms and architecures) do not get an unfair advantage), plus between 0 and 80 documents from VS.
There are 47 test documents, which are held fixed as we vary over the other section and the
number N of training VS documents. We compared 3 different models for this task. The
first model (M1) simply ignores documents from the additional section, and uses a HDP to
model the VS documents. It serves as a baseline. The second model (M2) uses a HDP mixture model, with one group per document, but lumping together training documents from
both sections. The third model (M3) takes a hierarchical approach and models each section
separately using a HDP mixture model, and places another DP prior over the common base
distributions for both submodels4 .
As we see in Figure 3 left, the more hierarchical approach of M3 performs best, with perplexity decreasing drastically with modest values of N , while M1 does worst for small N .
However with increasing N , M1 improves until it is competitive with M3 but M2 does
worst. This is because M2 lumps all the documents together, so is not able to differentiate
between the sections, as a result the influence of documents from the other section is unduly strong. This result confirms that the hierarchical approach to the transfer-of-learning
problem is a useful one, as it allows useful information to be transfered to a new task (here
the modeling of a new section), without the data from the previous tasks overwhelming
those in the new task.
We also looked at the performance of the M3 model on VS documents given specific other
sections. This is shown in Figure 3 right. As expected, the performance is worst given LT
(learning theory), and improves as we move to AA and AP (applications). In Table 1 we
show the topics pertinent to VS discovered by the M3 model. First we trained the model
on all documents from the other section. Then, keeping the assignments of words to topics
fixed in the other section, we introduced VS documents and the model decides to reuse
some topics from the other section, as well as create new ones. The topics reused by VS
documents confirm to our expectations of the overlap between VS and other sections.
as words occurring more than 4000 or less than 50 times in the documents. As sections differ over
the years, we assigned by hand the various sections to one of 9 prototypical sections: CS, NS, LT,
AA, IM, SP, VS, AP and CN.
4
Though we have only described the 2 layer HDP the 3 layer extension is straightforward. In
fact on our website http://www.cs.berkeley.edu/?ywteh/research/npbayes we have an implementation of the
general case where DPs are coupled hierarchically in a tree-structured model.
LT
AA
task
representation
pattern
processing
trained
representations
three process
unit patterns
CS
cells cell
activity
response
neuron visual
patterns
pattern single
fig
NS
signal layer
gaussian cells
fig nonlinearity
nonlinear rate
eq cell
algorithms test
approach
methods based
point problems
form large
paper
processing
pattern
approach
architecture
single shows
simple based
large control
IM
visual images
video language
image pixel
acoustic delta
lowpass flow
SP
approach
based trained
test layer
features table
classification
rate paper
AP
ii tree pomdp
observable
strategy class
stochastic
history
strategies
density
CN
examples
concept
similarity
bayesian
hypotheses
generalization
numbers
positive classes
hypothesis
visual cells
cortical
orientation
receptive
contrast spatial
cortex stimulus
tuning
large examples
form point see
parameter
consider
random small
optimal
distance
tangent image
images
transformation
transformations
pattern vectors
convolution
simard
motion visual
velocity flow
target chip eye
smooth
direction optical
signals
separation
signal sources
source matrix
blind mixing
gradient eq
image images
face similarity
pixel visual
database
matching facial
examples
policy optimal
reinforcement
control action
states actions
step problems
goal
Table 1: Topics shared between VS and the other sections. Shown are the two topics with most
numbers of VS words, but also with significant numbers of words from the other section.
Alice in Wonderland. The infinite hidden Markov model (iHMM) is a nonparametric
model for sequential data where the number of hidden states is open-ended and inferred
from data [11]. In [10] we show that the HDP framework can be applied to obtain a cleaner
formulation of the iHMM, providing effective new inference algorithms and potentially hierarchical extensions. In fact the original iHMM paper [11] served as inspiration for this
work and first coined the term ?hierarchical Dirichlet processes??though their model is
not hierarchical in the Bayesian sense, involving priors upon priors, but is rather a set of
coupled urn models similar to the CRF. Here we report experimental comparisons of the
iHMM against other approaches on sentences taken from Lewis Carroll?s Alice?s Adventures in Wonderland.
Perplexity on test sentences of Alice
50
ML
MAP
VB
Perplexity
40
30
20
10
0
0
5
10
15
20
Number of hidden states
25
30
Figure 4: Comparing iHMM (horizontal line)
versus ML, MAP and VB trained HMMs. Error bars are 1 standard error (those for iHMM too
small to see).
ML, MAP, and variational Bayesian (VB)
[12] models with numbers of states ranging from 1 to 30 were trained multiple
times on 20 sentences of average length
51 symbols (27 distinct symbols, consisting of 26 letters and ? ?), and tested on
40 sequences of average length 100. Figure 4 shows the perplexity of test sentences. For VB, the predictive probability
is intractable to compute, so the modal setting of parameters was used. Both MAP
and VB models were given optimal settings of the hyperparameters found in the
iHMM. We see that the iHMM has a lower
perlexity than every model size for ML,
MAP, and VB, and obtains this with one
countably infinite model.
6 Discussion
We have described the hierarchical Dirichlet process, a hierarchical, nonparametric model
for clustering problems involving multiple groups of data. HDP mixture models are able
to automatically determine the appropriate number of mixture components needed, and
exhibit sharing of statistical strength across groups by having components shared across
groups. We have described the HDP as a distribution over distributions, using both the
stick-breaking construction and the Chinese restaurant franchise. In [10] we also describe
a fourth perspective based on the infinite limit of finite mixture models, and give detail for
how the HDP can be applied to the iHMM. Direct extensions of the model include use of
nonparametric priors other than the DP, building higher level hierarchies as in our NIPS
experiment, as well as hierarchical extensions to the iHMM.
Appendix: Gibbs Sampling in the CRF
The CRF is defined by the variables t = (tji ), k = (kjt ), and ? = (?k ). We describe an
inference procedure for the HDP mixture model based on Gibbs sampling t, k and ? given
data items x. For the full derivation see [10]. Let f (?|?) and h be the density functions for
?jt
F (?) and H respectively, n?i
be
jt be the number of tji0 ?s equal to t except tji , and mk
the number of kj 0 t0 ?s equal to k except kjt . The conditional probability for tji given the
other variables is proportional to the product of a prior and likelihood term. The prior term
is given by (7) where, by exchangeability, we can take tji to be the last one assigned. The
likelihood is given by f (xji |?kjt ) where for t = tnew we may sample kjtnew using (8), and
?knew ? H. The distribution is then:
?0 f (xji |?kjt ) if t = tnew
p(tji = t | t\tji , k, ?, x) ?
(9)
n?i
jt f (xji |?kjt ) if t currently used.
Similarly the conditional distributions for kjt and ?k are:
( Q
? i:tji =t f (xji |?k )
if k = k new
p(kjt = k | t, k\kjt , ?, x) ?
?t Q
mk
i:tji =t f (xji |?k ) if k currently used.
Y
p(?k | t, k, ?\?k , x) ? h(?k )
f (xji |?k )
(10)
(11)
ji:kjtji =k
where ?knew ? H. If H is conjugate to F (?) we have the option of integrating out ?.
References
[1] D.M. Blei, A.Y. Ng, and M.I. Jordan. Latent Dirichlet allocation. JMLR, 3:993?1022, 2003.
[2] M.D. Escobar and M. West. Bayesian density estimation and inference using mixtures. Journal
of the American Statistical Association, 90:577?588, 1995.
[3] S.N. MacEachern and P. M?uller. Estimating mixture of Dirichlet process models. Journal of
Computational and Graphical Statistics, 7:223?238, 1998.
[4] T.S. Ferguson. A Bayesian analysis of some nonparametric problems. Annals of Statistics,
1(2):209?230, 1973.
?
[5] D. Aldous. Exchangeability and related topics. In Ecole
d??et?e de probabilit?es de Saint-Flour
XIII?1983, pages 1?198. Springer, Berlin, 1985.
[6] J. Sethuraman. A constructive definition of Dirichlet priors. Statistica Sinica, 4:639?650, 1994.
[7] R.M. Neal. Markov chain sampling methods for Dirichlet process mixture models. Journal of
Computational and Graphical Statistics, 9:249?265, 2000.
[8] C.E. Rasmussen. The infinite Gaussian mixture model. In NIPS, volume 12, 2000.
[9] D.M. Blei, T.L. Griffiths, M.I. Jordan, and J.B. Tenenbaum. Hierarchical topic models and the
nested Chinese restaurant process. NIPS, 2004.
[10] Y.W. Teh, M.I. Jordan, M.J. Beal, and D.M. Blei. Hierarchical dirichlet processes. Technical
Report 653, Department of Statistics, University of California at Berkeley, 2004.
[11] M.J. Beal, Z. Ghahramani, and C.E. Rasmussen. The infinite hidden Markov model. In NIPS,
volume 14, 2002.
[12] M.J. Beal. Variational Algorithms for Approximate Bayesian Inference. PhD thesis, Gatsby
Unit, University College London, 2004.
| 2698 |@word achievable:1 proportion:6 reused:1 open:3 confirms:1 covariance:1 ecole:1 document:27 comparing:1 must:1 subsequent:1 realistic:1 informative:1 partition:7 j1:3 pertinent:1 plot:1 v:18 implying:1 generative:2 discovering:1 website:1 item:1 blei:5 node:1 toronto:3 sits:2 unbounded:1 direct:1 beta:2 introduce:1 g4:1 expected:1 indeed:1 xji:8 nor:1 multi:2 decreasing:1 automatically:4 metaphor:2 overwhelming:1 increasing:1 begin:1 discover:1 underlying:1 provided:1 estimating:1 mass:2 differing:1 transformation:2 ended:3 nj:3 berkeley:5 every:1 stick:9 control:2 exchangeable:3 unit:3 partitioning:1 positive:4 understood:2 limit:1 mainstay:1 ap:6 plus:1 alice:3 ease:2 hmms:1 range:1 averaged:3 practical:1 atomic:2 practice:1 procedure:1 probabilit:1 empirical:1 matching:1 word:12 integrating:1 refers:2 griffith:1 get:2 marginalize:3 selection:1 applying:2 influence:1 yee:1 www:1 measurable:1 map:5 customer:15 straightforward:1 independently:4 pomdp:1 identifying:1 lumping:1 m2:4 submodel:1 menu:1 notion:4 analogous:1 annals:1 construction:6 play:1 imagine:1 suppose:1 target:1 hierarchy:1 us:2 hypothesis:2 associate:3 velocity:1 jk:6 database:1 role:1 capture:1 worst:3 trained:5 depend:1 segment:1 predictive:1 upon:2 vague:1 resolved:1 joint:1 lowpass:1 k0:3 chip:1 various:3 derivation:1 distinct:6 effective:2 describe:7 london:1 nematode:2 whose:1 larger:1 statistic:6 g1:1 itself:1 beal:5 differentiate:1 advantage:1 sequence:1 propose:1 product:1 tj1:1 mixing:7 description:1 cluster:14 requirement:1 extending:1 generating:2 escobar:1 franchise:3 help:1 eq:2 strong:1 c:5 implies:1 come:1 differ:1 direction:1 tji:11 subsequently:1 shrunk:1 stochastic:1 require:1 assign:3 generalization:5 im:2 extension:4 around:1 matthew:1 vary:2 estimation:1 bag:1 currently:3 kjt:12 stickbreaking:1 grouped:4 xjnj:1 create:1 tool:2 weighted:1 uller:1 gaussian:6 always:2 rather:2 occupied:1 exchangeability:3 improvement:1 likelihood:2 contrast:1 baseline:1 sense:1 inference:4 ferguson:1 integrated:2 hidden:5 selects:1 pixel:2 overall:1 among:5 classification:1 orientation:1 priori:1 plan:1 spatial:1 equal:2 once:1 having:1 ng:1 atom:3 sampling:4 biology:2 report:3 stimulus:1 xiii:1 gamma:1 individual:2 consisting:2 n1:1 wli:1 flour:1 introduces:1 mixture:44 held:2 chain:1 facial:1 modest:1 tree:2 mk:3 formalism:1 modeling:2 ar:4 assignment:2 uninteresting:1 too:1 dependency:1 dir:1 npbayes:1 density:3 grand:1 off:1 michael:1 together:2 transfered:1 again:2 thesis:1 american:1 simard:1 de:2 explicitly:1 blind:1 performed:1 view:2 start:1 competitive:1 option:1 qk:2 characteristic:1 who:1 yield:1 identify:1 sitting:1 correspond:4 prefering:1 bayesian:13 tnew:3 served:3 m5s:1 history:1 sharing:4 definition:4 against:2 ihmm:10 associated:1 couple:1 stop:2 sampled:1 dataset:1 begun:1 popular:1 recall:1 realm:1 infers:1 improves:2 organized:2 higher:1 response:1 modal:1 formulation:1 though:2 just:1 crp:6 until:1 hand:2 horizontal:1 unoccupied:1 nonlinear:1 defines:1 lda:8 grows:2 usa:1 building:2 xj1:1 concept:1 hence:2 assigned:2 inspiration:1 symmetric:1 neal:1 conditionally:2 swmed:1 whye:1 complete:1 crf:6 demonstrate:1 performs:2 motion:1 image:6 adventure:1 variational:2 ranging:1 superior:1 common:3 ji:12 overview:1 volume:2 jl:1 extend:1 association:1 m1:4 interpret:1 wonderland:2 significant:2 refer:1 gibbs:4 smoothness:1 tuning:1 similarly:1 nonlinearity:1 language:1 had:1 similarity:2 cortex:1 gj:17 carroll:1 base:7 posterior:6 perspective:2 aldous:1 dish:3 perplexity:12 seen:1 additional:5 determine:1 paradigm:2 signal:3 ii:1 multiple:8 full:1 smooth:3 technical:3 a1:4 ensuring:1 involving:5 basic:1 vision:1 expectation:1 histogram:1 cell:5 want:1 separately:1 completes:1 source:2 elegant:1 elegans:1 flow:2 jordan:5 call:2 integer:2 lump:1 variety:1 restaurant:13 architecture:1 idea:2 cn:2 t0:1 motivated:2 reuse:1 jj:1 repeatedly:1 action:2 ignored:1 generally:1 useful:4 detailed:1 cleaner:1 nonparametric:13 tenenbaum:1 induces:1 generate:2 http:2 notice:2 delta:1 hdps:1 per:1 write:2 discrete:1 group:32 demonstrating:1 drawn:3 sum:1 year:1 run:3 letter:1 powerful:1 fourth:1 place:2 separation:1 putative:1 draw:11 appendix:2 prefer:1 vb:6 layer:4 conferring:1 activity:1 strength:4 occur:1 n3:1 flat:1 ywteh:2 aspect:2 performing:1 urn:1 optical:1 structured:1 department:1 according:2 combination:1 conjugate:1 across:11 g3:1 taken:1 previously:1 turn:1 discus:1 describing:1 needed:1 know:2 serf:1 available:1 apply:2 eight:1 hierarchical:23 appropriate:1 appearing:1 frequentist:1 alternative:1 jnj:3 original:1 dirichlet:19 clustering:11 ensure:1 include:1 graphical:3 saint:1 coined:1 ghahramani:1 chinese:10 classical:1 move:1 g0:36 already:1 looked:1 parametric:1 concentration:4 strategy:2 receptive:1 div:1 exhibit:2 dp:35 gradient:1 distance:1 berlin:1 hmm:1 topic:23 seating:1 hdp:39 length:4 modeled:3 relationship:1 providing:1 culled:1 setup:3 unfortunately:1 sinica:1 potentially:1 implementation:1 policy:1 models2:1 teh:2 allowing:1 observation:1 neuron:1 markov:4 convolution:1 finite:4 relational:1 precise:1 discovered:1 varied:1 canada:1 inferred:2 david:1 introduced:1 sentence:4 conflict:1 california:2 acoustic:1 unduly:1 nip:8 able:3 bar:3 exemplified:1 pattern:6 including:1 video:1 overlap:1 x2i:1 brief:1 eye:2 sethuraman:1 coupled:2 kj:1 text:2 prior:10 discovery:1 tangent:1 highlight:2 prototypical:1 proportional:3 allocation:2 versus:1 consistent:1 njt:2 share:7 last:1 keeping:1 rasmussen:2 drastically:1 allow:1 understand:1 taking:2 face:1 distributed:4 curve:1 xn:1 evaluating:1 vocabulary:1 cortical:1 ignores:1 made:1 reinforcement:1 approximate:1 observable:1 obtains:1 countably:2 dealing:1 confirm:1 global:4 decides:1 reveals:1 instantiation:1 sat:2 corpus:7 conceptual:1 assumed:2 ml:4 knew:3 xi:1 latent:3 table:18 learn:1 nature:6 transfer:2 ca:1 sp:2 pk:1 hierarchically:1 statistica:1 hyperparameters:1 n2:1 x1:1 fig:2 referred:1 west:1 gatsby:1 n:2 wish:1 explicit:1 x1i:1 unfair:1 breaking:9 jmlr:1 third:1 down:2 removing:1 specific:2 jt:11 showing:1 k21:1 symbol:2 sit:3 k11:1 consist:1 intractable:1 sequential:1 effectively:1 phd:1 occurring:1 logarithmic:1 lt:6 simply:4 x3i:1 visual:5 expressed:1 g2:1 applies:1 springer:1 aa:7 corresponds:3 nested:1 lewis:1 conditional:4 goal:3 viewed:2 towards:2 shared:6 infinite:6 specifically:2 typical:1 except:3 sampler:1 total:2 experimental:3 e:1 m3:7 kjtji:1 college:1 maceachern:1 constructive:1 dept:2 tested:1 |
1,870 | 2,699 | Optimal Aggregation of Classifiers and Boosting
Maps in Functional Magnetic Resonance
Imaging
Vladimir Koltchinskii
Department of Mathematics and Statistics
University of New Mexico
Albuquerque, NM, 87131
Manel Mart??nez-Ram?on
Department of Electrical and Computer Engineering
University of New Mexico
Albuquerque, NM, 87131
Stefan Posse
Department of Psychiatry and The Mind Institute
University of New Mexico
Albuquerque, NM, 87131
Abstract
We study a method of optimal data-driven aggregation of classifiers in a
convex combination and establish tight upper bounds on its excess risk
with respect to a convex loss function under the assumption that the solution of optimal aggregation problem is sparse. We use a boosting type
algorithm of optimal aggregation to develop aggregate classifiers of activation patterns in fMRI based on locally trained SVM classifiers. The
aggregation coefficients are then used to design a ?boosting map? of the
brain needed to identify the regions with most significant impact on classification.
1
Introduction
We consider a problem of optimal aggregation (see [1]) of a finite set of base classifiers in
a complex aggregate classifier. The aggregate classifiers we study are convex combinations
of base classifiers and we are using boosting type algorithms as aggregation tools. Building
upon recent developments in learning theory, we show that such boosting type aggregation
yields a classifier with a small value of excess risk in the case when optimal aggregate
classifiers are sparse and that, moreover, the procedure provides reasonably good estimates
of aggregation coefficients. Our primary goal is to use this approach in the problem of
classification of activation patterns in functional Magnetic Resonance Imaging (fMRI) (see,
e.g., [2]).
In these problems it is of interest not only to classify the patterns, but also to determine
areas of the brain that are relevant for a particular classification task. Our approach is based
on splitting the image into a number of functional areas, training base classifiers locally in
each area and then combining them into a complex aggregate classifier. The aggregation
coefficients are used to create a special representation of the image we call the boosting
map of the brain. It is needed to identify the functional areas with the most significant
impact on classification.
Previous work has focused on classifying patterns within subject [2] and these patterns were located in the occipital lobe. Here we are considering a different problem, that
is widely distributed patterns in multiple brain regions across groups of subjects. We use
prior knowledge from functional neuroanatomical brain atlases to subdivide the brain into
Regions of Interest, which makes this problem amenable to boosting. Classification across
subjects requires spatial normalization to account for inter-subject differences in brain size
and shape, but also needs to be robust with respect to inter-subject differences in activation
patterns ?shape and amplitude.
Since fMRI patterns are very high dimensional and the amount of training data is
typically limited, some form of ?bet on sparsity? principle (?use a procedure that does well
in sparse problems, since no procedure does well in dense problems? see [3]) becomes
almost unavoidable and our theoretical analysis shows that boosting maps might have a
good chance of success in sparse problems (when only few functional areas are relevant for
classification).
2
Optimal aggregation of classifiers
Although we developed a multiclass extension of the method, for simplicity, we are dealing
here with a standard binary classification. Let (X, Y ) be a random couple with distribution
P, X being an instance in some space S (e.g., it might be an fMRI pattern) and Y ?
{?1, 1} being a binary label. Here and in what follows all the random variables are defined
on a probability space (?, ?, P), E denotes the expectation. Functions f : S 7? R will
be used as classifiers, sign(f (x)) being a predictor of the label for an instance x ? S (no
decision is being made if f (x) = 0). The quantity P {(x, y) : yf (x) ? 0} (the probability
of misclassification or abstaining) is called the generalization error or the risk of f. Suppose
that H := {h1 , . . . , hN } is a given family of classifiers taking values in [?1, 1]. Let
X
N
N
X
|?j | ? 1
conv(H) :=
? j hj :
j=1
j=1
be the symmetric convex hull of H. One of the versions of optimal aggregation problem
would be to find a convex combination f ? conv(H) that minimizes the generalization
error of f in conv(H). For a given f ? conv(H) its quality is measured by
E(f ) := P {(x, y) : yf (x) ? 0} ?
inf
g?conv(H)
P {(x, y) : yg(x) ? 0},
which is often called the excess risk of f. Since the true distribution P of (X, Y ) is unknown, the solution of the optimal aggregation problem is to be found based on the training
data (X1 , Y1 ), . . . , (Xn , Yn ) consisting of n independent copies of (X, Y ).
Let Pn denote the empirical measure based on the training data, i.e., Pn (A) represents
the frequency of training examples in a set A ? S?{?1, 1}. In what follows, we denote P h
or Pn h the integrals of a function h on S ? {?1, 1} with respect to P or Pn , respectively.
We use the same notation for functions on S with an obvious meaning.
Since the generalization error is not known, it is tempting to try to estimate the optimal
convex aggregate classifier by minimizing the training error Pn {(x, y) : yf (x) ? 0}
over the convex hull conv(H). However, this minimization problem is not computationally
feasible and, moreover, the accuracy of empirical approximation (approximation of P by
Pn ) over the class of sets {{(x, y) : yf (x) ? 0} : f ? conv(H)} is not good enough
when H is a large class. An approach that allows one to overcome both difficulties and that
proved to be very successful in the recent years is to replace the minimization of the training
error by the minimization of the empirical risk with respect to a convex loss function. To
be specific, let ? be a nonnegative decreasing convex function on R such that ?(u) ? 1 for
u ? 0. We will denote (? ? f )(x, y) := ?(yf (x)). The quantity
Z
P (? ? f ) = (? ? f )dP = E?(Y f (X))
is called the risk of f with respect to the loss ?, or the ?-risk of f. We will call a function
f0 :=
N
X
?0j hj ? conv(H)
j=1
an ?-otimal aggregate classifier if it minimizes the ?-risk over conv(H). Similarly to the
excess risk, one can define the excess ?-risk of f as
E? (f ) := P (? ? f ) ?
inf
g?conv(H)
P (? ? g).
Despite the fact that we concentrate in what follows on optimizing the excess ?-risk
(?-optimal aggregation) it often provides also a reasonably good solution of the problem
of minimizing the generalization error (optimal aggregation), as it follows from simple
inequalities relating the two risks and proved in [4].
As before, since P is unknown, the minimization of ?-risk has to be replaced by the
corresponding empirical risk minimization problem
Pn (? ? f ) =
n
1X
? Yj f (Xj ) ?? min, f ? conv(H),
n i=1
PN ?
whose solution f? := j=1 ?
j hj is called an empirical ?-optimal aggregate classifier.
? j are small for most of the values
We will show that if f0 , f? are ?sparse? (i.e., ?0j , ?
of j), then the excess ?-risk of the empirical ?-optimal aggregate classifier is small and,
moreover, the coefficients of f? are close to the coefficients of f0 in ?1 -distance.
The sparsity assumption is almost unavoidable in many problems because of the ?bet
on sparsity? principle (see the Introduction).
At a more formal level, if there exists a small subset J ? {1, 2, . . . , N } such that the
sets of random variables {Y, hj (X), j ? J} and {hj (X), j 6? J} are independent and, in
addition, Ehj (X) = 0, j 6? J, then, using Jensen?s inequality, it is easy to check that in an
?-optimal aggregate classifier f0 one can take ?0j = 0, j 6? J.
PN
We will define a measure of sparsity of a function f := j=1 ?j hj ? conv(H) that
is somewhat akin to sparsity charactersitics considered in [5, 6]. For 0 ? d ? N, let
X
?(f ; d) := min
|?j | : J ? {1, . . . , N }, card(J) = d
j6?J
and let ?n (d) :=
Define
d log(N n/d)
.
n
p
dn (f ) := min d : 1 ? d ? N, ?n (d) ? ?(d) .
Of course, if there exists J ? {1, . . . , N } such that ?j = 0 for all j 6? J and
card(J) = d, then dn (f ) ? d.
We will also need the following measure of linear independence of functions in H :
?1
X
?(d) := ?(H; d) =
inf
?j hj
.
P inf
J?{1,...,N },card(J)=d
j?J
|?j |=1
j?J
L2 (P )
Finally, we need some standard conditions on the loss function ? (as, for instance, in
[4]). Assume that ? is Lipschitz on [?1, 1] with some constant L, |?(u) ? ?(v)| ? L|u ?
v|, u, v ? [?1, 1], and the following condition on the convexity modulus of ? holds with
??L:
?(u) + ?(v)
u+v
??
? ?|u ? v|2 , u, v ? [?1, 1].
2
2
In fact, ?(u) is often replaced by a function ?(uM ) with a large enough M (in other
words, the ?-risk is minimized over M conv(H)). This is the case, for instance, for so called
regularized boosting [7]. The theorem below applies to this case as well, only a simple
rescaling of the constants is needed.
Theorem 1 There exist constants K1 , K2 > 0 such that for all t > 0
r
^ log N
t
L2
?
?
?n (dn (f ))
+
? e?t
P E? (f ) ? K1
?
n
n
and
r
X
N
L
0
? j ? ? | ? K2 ?(dn (f?) + dn (f0 )) ?n (dn (f?) + dn (f0 )) + t ? e?t .
P
|?
j
?
n
j=1
Our proof requires some background material on localized Rademacher complexities
and their role in bounding of excess risk (see [8]). We defer it to the full version of the
paper. Note that the first bound depends only on dn (f?) and the second on dn (f?), dn (f0 ).
Both quantities can be much smaller than N despite the fact that empirical risk minimization occurs over the whole N -dimensional convex hull. However, the approach to convex
aggregation based on minimization of the empirical ?-risk over the convex hull does not
guarantee that f? is sparse even if f0 is. To address this problem, we also studied another
approach based on minimization of the penalized empirical ?-risk with the penalty based
on the number of nonzero coefficients of the classifier, but the size of the paper does not
allow us to discuss it.
3
Classification of fMRI patterns and boosting maps
We are using optimal aggregation methods described above in the problem of classification
of activation patterns in fMRI. Our approach is based on dividing the training data into
two parts: for local training and for aggregation. Then, we split the image into N functional areas and train N local classifiers h1 , . . . , hN based on the portions of fMRI data
corresponding to the areas. The data reserved for aggregation is then used to construct an
aggregate classifier. In applications, we are often replacing direct minimization of empirical
risk with convex loss by the standard AdaBoost algorithm (see, e.g., [9]), which essentially
means choosing the loss function as ?(u) = e?u . A weak (base) learner for AdaBoost simply chooses in this case a local classifier among h1 , . . . , hN with the smallest weighted
training error [in more sophisticated versions, we choose a local classifier at random with
probability depending on the size of its weighted training error] and after a number of
rounds AdaBoost returns a convex combination of local classifiers. The coefficients of this
aggregate classifier are then used to create a new visual representation of the brain (the
boosting map) that highlights the functional areas with significant impact on classification.
In principle, it is also possible to use the same data for training of local classifiers and for
aggregation (retraining the local classifiers at each round of boosting), but this approach is
time consuming.
We use statistical parametric model (SPM) t-maps of MRI scans [10]. Statistical parametric maps (SPMs) are image processes with voxel1 values that are, under the null hypothesis, distributed according to a known probability density function, usually the Student?s
1
A voxel is the amplitude of a position in the 3-D MRI image matrix.
Figure 1: Masks used to split the image into functional areas in multi-slice and 3 orthogonal
slice display representations.
T or F distributions. These are known colloquially as t- or f-maps. Namely, one analyzes
each and every voxel using any standard (univariate) statistical test. The resulting statistical
parameters are assembled into an image - the SPM.
The classification system essentially transforms the t-map of the image into the boosting map and at the same time it returns the aggregate classifier. The system consists of the
data preprocessing block that splits the image into functional areas based on specified
masks, and also splits the data into portions corresponding to the areas. In one of our examples, we use the main functional areas brainstem, cerebellum, occipital, temporal, parietal,
subcortical and frontal. We split these masks in left and right, having in total 14 of them.
The classifier block then trains local classifiers based on local data (in the current version
we are using SVM classifiers). Finally, the aggregation or boosting block computes and
outputs the aggregate classifier and the boosting map of the image. We developed a version of the system that deals with multi-class problems in spirit of [11], but the details go
beyond the scope of this paper. The architecture of the network allows us also to train it
sequentially. Let f be a classifier produced by the network in the previous round of work,
let (X1 , Y1 ), . . . , (Xn , Yn ) be either the same or a new training data set and let h1 , . . . , hN
be local classifiers (based either on the same, or on a new set of masks). Then one can
?Yj f (Xj )
, where Z is a standard
assign to the training examples the initial weights wj = e Z
normalizing constant, instead of usually chosen uniform weights. After this, the AdaBoost
can proceed in a normal fashion creating at the end an aggregate of f and of new local classifiers. The process can be repeated recursively updating both the classifier and the boosting
map.
5
5
5
10
10
10
15
15
15
20
20
20
25
25
25
30
30
30
35
35
5
10
15
20
25
30
35
35
5
10
15
20
25
30
35
5
10
15
20
25
30
35
Figure 2: Left and center: Patterns corresponding to two classes of data. Right: Locations
of the learners chosen by the boosting procedure (white spots). The background image
corresponds to the two patterns of left and center figures superimposed.
5
5
10
10
15
15
20
20
25
25
30
30
35
35
5
10
15
20
25
30
35
5
10
15
20
25
30
35
Figure 3: Patterns corrupted with noise in the gaussian parameters, artifacts, and additive
noise used in the synthetic data experiment.
Figure 4: Two t-maps corresponding to visual (left) and motor activations in the same
subject used in the real data experiment.
As a synthetic data example, we generate 40 ? 40 pixels images of two classes. Each
class of images consists of three gaussian clusters placed in different positions. We generate
the set of images by adding gaussian noise of standard deviation 0.1 to the standard deviation and position of the clusters. Then, we add 10 more clusters with random parameters,
and finally, additive noise of standard deviation 0.1. Figure 2 (left and center) shows the
averages of class 1 and class 2 images respectively. Two samples of the images can be seen
in Figure 3
We apply a base learner to each one of the 1600 pixels of the images. Learners have
been trained with 200 data, 100 of each class, and the aggregation has been trained with
200 more data. The classifier has been tested with 200 previously unknown data. The error
averaged over 100 trials is of 9.5%. The same experiment has been made with a single linear
SVM, producing an error which exceeds 20%, although this rate can be slightly improved
by selecting C by cross validation.
The resulting boosting map can be seen in Fig. 2 (right). As a proof of concept, we
remark that the map is able to focus in the areas in which the clusters corresponding to each
class are, discarding those areas in which only randomly placed clusters are present.
In order to test the algorithm in a real fMRI experiment, we use 20 images taken from
10 healthy subjects on a 1.5 Tesla Siemens Sonata scanner. Stimuli were presented via MR
compatible LCD goggles and headphones. The paradigm consists of four interleaved tasks:
visual (8 Hz checkerboard stimulation), motor (2 Hz right index finger tapping), auditory
Figure 5: Boosting map of the brain corresponding to the classification problem with visual
and motor activations. Darker regions correspond to higher values.
left brainstem:
left cerebellum:
left parietal:
left temporal:
left occipital:
left subcortical:
left frontal:
0
0.15
0.02
0.03
0.29
0
0
right brainstem:
right cerebellum:
right parietal:
right temporal:
right occipital:
right subcortical:
right frontal:
0
0.16
0.06
0.15
0.15
0
0
Table 1: Values of the convex aggregation.
(syllable discrimination) and cognitive (mental calculation). These tasks are arranged in
randomized blocks (8 s per block). Finger tapping in the motor task was regulated with
an auditory tone, subjects were asked to tap onto a button-response pad. During the auditory task, subjects were asked to respond on a button-response pad for each ?Ta? (25% of
sounds), but not to similar syllables. Mental calculation stimuli consisted of three singledigit numbers heard via headphone. Participants had to sum them and divide by three,
responding by button press when there was no remainder (50% of trials).
Functional MRI data were acquired using single-shot echo-planar imaging with TR:
2 s, TE: 50 ms, flip angle: 90 degrees, matrix size: 64 ? 64 pixels, FOV: 192 mm. Slices
were 6 mm thick, with 25% gap, 66 volumes were collected for a total measurement time
of 132 sec per run. Statistical parametric mapping was performed to generate t-maps that
represent brain activation changes.
The t-maps are lowpass filtered and undersampled to obtain 32 ? 32 ? 24 t-maps (Fig.
4). The resulting t-maps are masked to obtain 14 subimages, then the data is normalized
in amplitude. We proceed as mentioned to train a set of 14 Support Vector Machines. The
used kernel is a gaussian one with ? = 2 and C = 10. These parameters have been chosen
to provide an acceptable generalization. A convex aggregation of the classifier outputs is
then trained.
We tested the algorithm in binary classification of visual against auditory activations.
We train the base learners with 10 images, and the boosting with 9. Then, we train the
base learners again with 19, leaving one for testing. We repeat the experiment leaving
out a different image each trial. None of the images was misclassified. The values for the
aggregation are in Table 1. The corresponding boosting map is shown in Fig 5. It highlights
the right temporal and both occipital areas, where the motor and visual activations are
present (see Fig. 4). Also, there is activation in the cerebellum area in some of the motor
t-maps, which is highlighted by the boosting map.
In experiments for the six binary combination of activation stimuli, the average error
was less than 10%. This is an acceptable result if we take into account that the data included
ten different subjects, whose brain activation patterns present noticeable differences.
4
Future goals
Boosting maps we introduced in this paper might become a useful tool in solving classification problems for fMRI data, but there is a number of questions to be answered before it
is the case. The most difficult problem is the choice of functional areas and local classifiers
so that the ?true? boosting map is identifiable based on the data. As our theoretical analysis
shows, this is related to the degree of linear independence of local classifiers quantified by
the function ?(d). If ?(d) is too large for d = dn (f0 ) ? dn (f?), the empirical boosting map
can become very unstable and misleading. In such cases, there is a challenging model selection problem (how to choose a ?good? subset of H or how to split H into ?almost linearly
independent clusters? of functions) that has to be addressed to develop this methodology
further.
Acknowledgments
We want to acknowledge to Jeremy Bockholt (MIND Institute) for providing the brain
masks, generated with BRAINS2. Partially supported by NSF Grant DMS-0304861 and
NIH Grant NIBIB 1 RO1 EB002618-01, Dept. of Mathematics and Statistics, Dept. of
Electrical and Computing Engineering, Dept. of Psychiatry and The MIND Institute.
References
[1] Tsybakov, A. (2003) Optimal rates of aggregation. In: COLT2003, Lecture Notes in Artificial
Intelligence, Eds.: M. Warmuth and B. Schoelkopf, Springer.
[2] Cox, D.D., Savoy, R.L. (2003) Functional magnetic resonance imaging (fMRI) ?brain reading?: detecting and classifying distributed patterns of fMRI activity in human visual cortex,
Neuroimage19, 2, 261?70.
[3] Friedman, J., Hastie, T., Rosset, S., Tibshirani, R. and Zhu, J. (2004) Discussion on Boosting,
Annals of Statistics, 32, 1, 102?107.
[4] Bartlett, P. L., Jordan, M.I., McAuliffe, J. D. (2003) Convexity, classification, and risk
bounds. Technical Report 638, Department of Statistics, U.C. Berkeley, 2003. Journal of
the American Statistical Association.To appear.
[5] Koltchinskii, V., Panchenko, D. and Lozano, F. (2003) Bounding the generalization error of
combined classifiers: balancing the dimensionality and the margins. A. Appl. Prob. , 13, 1.
[6] Koltchinskii, V., Panchenko, D. and Andonova, S. (2003) Generalization bounds for voting classifiers based on sparsity and clustering. In: COLT2003, Lecture Notes in Artificial
Intelligence, Eds.: M. Warmuth and B. Schoelkopf, Springer.
[7] Blanchard, G., Lugosi, G. and Vayatis, N. (2003) On the rates of convergence of regularized
boosting classifiers. Journal of Machine Learning Research 4, 861-894.
[8] Koltchinskii, V. (2003) Local Rademacher Complexities and Oracle Inequalities in Risk
Minimization. Preprint.
[9] Schapire, R. E. (1999) A brief Introduction to Boosting. In: Proc. of the 6th Intl. Conf. on
Artificial Inteligence.
[10] Friston, K., Frith, C., Liddle, P. and Frackowiak, R. (1991) Comparing functional (PET)
images: the assessment of significant change. J. Cereb. Blood Flow Met.11, 690-699
[11] Allwein, E. L., Schapire, R. E., and Singer, Y. (2000) Reducing multiclass to binary: A
unifying approach for margin classifiers. J. Machine Learning Research, 1, 113-141.
| 2699 |@word trial:3 cox:1 mri:3 version:5 retraining:1 lobe:1 tr:1 shot:1 recursively:1 initial:1 selecting:1 current:1 comparing:1 activation:12 additive:2 shape:2 motor:6 atlas:1 discrimination:1 intelligence:2 tone:1 warmuth:2 filtered:1 mental:2 provides:2 boosting:28 detecting:1 location:1 dn:12 direct:1 become:2 consists:3 acquired:1 inter:2 mask:5 multi:2 brain:13 decreasing:1 considering:1 becomes:1 conv:13 colloquially:1 moreover:3 notation:1 null:1 what:3 minimizes:2 developed:2 guarantee:1 temporal:4 berkeley:1 every:1 voting:1 um:1 classifier:46 k2:2 grant:2 yn:2 producing:1 mcauliffe:1 appear:1 before:2 engineering:2 local:14 despite:2 tapping:2 lugosi:1 might:3 koltchinskii:4 studied:1 quantified:1 fov:1 challenging:1 appl:1 limited:1 averaged:1 acknowledgment:1 yj:2 testing:1 block:5 spot:1 procedure:4 area:17 empirical:11 word:1 onto:1 close:1 selection:1 risk:23 map:27 center:3 go:1 occipital:5 convex:16 focused:1 simplicity:1 splitting:1 manel:1 annals:1 suppose:1 hypothesis:1 located:1 updating:1 role:1 preprint:1 electrical:2 region:4 wj:1 schoelkopf:2 mentioned:1 panchenko:2 convexity:2 complexity:2 asked:2 lcd:1 trained:4 tight:1 solving:1 upon:1 learner:6 lowpass:1 frackowiak:1 finger:2 train:6 artificial:3 aggregate:15 choosing:1 whose:2 widely:1 statistic:4 highlighted:1 echo:1 remainder:1 relevant:2 combining:1 convergence:1 cluster:6 rademacher:2 intl:1 depending:1 develop:2 measured:1 noticeable:1 dividing:1 met:1 concentrate:1 thick:1 hull:4 brainstem:3 human:1 material:1 assign:1 generalization:7 extension:1 hold:1 scanner:1 mm:2 considered:1 normal:1 scope:1 mapping:1 smallest:1 proc:1 label:2 healthy:1 create:2 tool:2 weighted:2 stefan:1 minimization:10 gaussian:4 pn:9 hj:7 bet:2 allwein:1 focus:1 check:1 superimposed:1 psychiatry:2 typically:1 pad:2 misclassified:1 pixel:3 classification:15 among:1 development:1 resonance:3 spatial:1 special:1 construct:1 having:1 represents:1 fmri:11 minimized:1 future:1 stimulus:3 report:1 few:1 randomly:1 replaced:2 consisting:1 friedman:1 interest:2 headphone:2 goggles:1 amenable:1 integral:1 orthogonal:1 divide:1 theoretical:2 instance:4 classify:1 deviation:3 subset:2 predictor:1 uniform:1 masked:1 successful:1 too:1 corrupted:1 synthetic:2 chooses:1 rosset:1 combined:1 density:1 randomized:1 yg:1 again:1 nm:3 unavoidable:2 hn:4 choose:2 cognitive:1 creating:1 american:1 conf:1 rescaling:1 return:2 checkerboard:1 account:2 jeremy:1 student:1 sec:1 coefficient:7 blanchard:1 depends:1 performed:1 h1:4 try:1 portion:2 aggregation:26 participant:1 defer:1 accuracy:1 reserved:1 yield:1 identify:2 correspond:1 weak:1 albuquerque:3 produced:1 none:1 j6:1 ed:2 against:1 frequency:1 obvious:1 dm:1 proof:2 couple:1 auditory:4 proved:2 knowledge:1 dimensionality:1 amplitude:3 sophisticated:1 higher:1 ta:1 adaboost:4 response:2 improved:1 planar:1 methodology:1 arranged:1 replacing:1 assessment:1 spm:2 yf:5 artifact:1 quality:1 modulus:1 building:1 liddle:1 concept:1 true:2 consisted:1 normalized:1 lozano:1 symmetric:1 nonzero:1 deal:1 white:1 ehj:1 round:3 cerebellum:4 during:1 m:1 cereb:1 image:22 meaning:1 nih:1 functional:15 stimulation:1 volume:1 association:1 relating:1 significant:4 measurement:1 mathematics:2 similarly:1 had:1 f0:9 cortex:1 base:7 add:1 recent:2 optimizing:1 inf:4 driven:1 inequality:3 binary:5 success:1 seen:2 analyzes:1 somewhat:1 mr:1 determine:1 paradigm:1 tempting:1 multiple:1 full:1 sound:1 exceeds:1 technical:1 calculation:2 cross:1 impact:3 essentially:2 expectation:1 normalization:1 represent:1 kernel:1 vayatis:1 addition:1 background:2 want:1 addressed:1 leaving:2 posse:1 subject:10 hz:2 flow:1 spirit:1 jordan:1 call:2 split:6 enough:2 easy:1 xj:2 independence:2 architecture:1 hastie:1 multiclass:2 six:1 bartlett:1 akin:1 penalty:1 proceed:2 remark:1 useful:1 heard:1 amount:1 transforms:1 tsybakov:1 locally:2 ten:1 generate:3 schapire:2 inteligence:1 exist:1 nsf:1 sign:1 per:2 tibshirani:1 group:1 four:1 blood:1 abstaining:1 imaging:4 ram:1 button:3 year:1 sum:1 run:1 angle:1 prob:1 respond:1 almost:3 family:1 decision:1 acceptable:2 interleaved:1 bound:4 syllable:2 display:1 nonnegative:1 identifiable:1 activity:1 oracle:1 answered:1 min:3 department:4 according:1 combination:5 across:2 smaller:1 slightly:1 ro1:1 taken:1 computationally:1 previously:1 discus:1 needed:3 mind:3 flip:1 singer:1 end:1 apply:1 magnetic:3 subdivide:1 neuroanatomical:1 denotes:1 responding:1 clustering:1 unifying:1 k1:2 establish:1 question:1 quantity:3 occurs:1 parametric:3 primary:1 regulated:1 dp:1 distance:1 card:3 collected:1 unstable:1 pet:1 index:1 providing:1 minimizing:2 vladimir:1 mexico:3 difficult:1 design:1 unknown:3 upper:1 finite:1 acknowledge:1 parietal:3 y1:2 introduced:1 namely:1 specified:1 tap:1 assembled:1 address:1 beyond:1 able:1 below:1 pattern:16 usually:2 sparsity:6 reading:1 misclassification:1 difficulty:1 friston:1 regularized:2 undersampled:1 zhu:1 misleading:1 brief:1 prior:1 l2:2 loss:6 lecture:2 highlight:2 subcortical:3 localized:1 validation:1 degree:2 principle:3 classifying:2 balancing:1 course:1 penalized:1 compatible:1 placed:2 repeat:1 copy:1 supported:1 formal:1 allow:1 institute:3 taking:1 sparse:6 distributed:3 slice:3 overcome:1 xn:2 computes:1 made:2 preprocessing:1 voxel:2 excess:8 nibib:1 dealing:1 sequentially:1 consuming:1 table:2 reasonably:2 robust:1 frith:1 complex:2 dense:1 main:1 linearly:1 bounding:2 whole:1 noise:4 repeated:1 tesla:1 x1:2 fig:4 fashion:1 darker:1 position:3 theorem:2 specific:1 discarding:1 jensen:1 svm:3 normalizing:1 exists:2 adding:1 subimages:1 te:1 margin:2 gap:1 nez:1 simply:1 univariate:1 visual:7 partially:1 applies:1 springer:2 corresponds:1 chance:1 mart:1 goal:2 replace:1 lipschitz:1 feasible:1 change:2 included:1 reducing:1 called:5 total:2 siemens:1 support:1 scan:1 frontal:3 dept:3 savoy:1 tested:2 |
1,871 | 27 | 573
BIT - SERIAL NEURAL NETWORKS
Alan F. Murray, Anthony V . W. Smith and Zoe F. Butler.
Department of Electrical Engineering, University of Edinburgh,
The King's Buildings, Mayfield Road, Edinburgh,
Scotland, EH93JL.
ABSTRACT
A bit - serial VLSI neural network is described from an initial architecture for a
synapse array through to silicon layout and board design. The issues surrounding bit
- serial computation, and analog/digital arithmetic are discussed and the parallel
development of a hybrid analog/digital neural network is outlined. Learning and
recall capabilities are reported for the bit - serial network along with a projected
specification for a 64 - neuron, bit - serial board operating at 20 MHz. This technique is extended to a 256 (2562 synapses) network with an update time of 3ms,
using a "paging" technique to time - multiplex calculations through the synapse
array.
1. INTRODUCTION
The functions a synthetic neural network may aspire to mimic are the ability to consider many solutions simultaneously, an ability to work with corrupted data and a
natural fault tolerance. This arises from the parallelism and distributed knowledge
representation which gives rise to gentle degradation as faults appear. These functions are attractive to implementation in VLSI and WSI. For example, the natural
fault - tolerance could be useful in silicon wafers with imperfect yield, where the
network degradation is approximately proportional to the non-functioning silicon
area.
To cast neural networks in engineering language, a neuron is a state machine that is
either "on" or "off', which in general assumes intermediate states as it switches
smoothly between these extrema. The synapses weighting the signals from a
transmitting neuron such that it is more or less excitatory or inhibitory to the receiving neuron. The set of synaptic weights determines the stable states and represents
the learned information in a system.
The neural state, VI' is related to the total neural activity stimulated by inputs to
the neuron through an activation junction, F. Neural activity is the level of excitation of the neuron and the activation is the way it reacts in a response to a change
in activation. The neural output state at time t, V[, is related to x[ by
V[
= F (xf)
(1)
The activation function is a "squashing" function ensuring that (say) Vi is 1 when
is large and -1 when Xi is small. The neural update function is therefore straightforward:
Xi
.
,
XI,+1 - XI
? ????
+ 0~
i-n-l
~
~
T ii V'J
(2)
J-O
where 8 represents the rate of change of neural activity, Tij is the synaptic weight
and n is the number of terms giving an n - neuron array [1].
Although the neural function is simple enough, in a totally interconnected n - neuron network there are n 2 synapses requiring n 2 multiplications and summations and
? American Institute of Physics 1988
574
a large number of interconnects. The challenge in VLSI is therefore to design a simple, compact synapse that can be repeated to build a VLSI neural network with
manageable interconnect. In a network with fixed functionality, this is relatively
straightforward. H the network is to be able to learn, however, the synaptic weights
must be programmable, and therefore more complicated.
2. DESIGNING A NEURAL NETWORK IN VLSI
There are fundamentally two approaches to implementing any function in silicon digital and analog. Each technique has its advantages and disadvantages, and these
are listed below, along with the merits and demerits of bit - serial architectures in
digital (synchronous) systems.
Digital vs. analog: The primary advantage of digital design for a synapse array is
that digital memory is well understood, and can be incorporated easily. Learning
networks are therefore possible without recourse to unusual techniques or technologies. Other strengths of a digital approach are that design techniques are advanced,
automated and well understood and noise immunity and computational speed can
be high. Unattractive features are that digital circuits of this complexity need to be
synchronous and all states and activities are quantised, while real neural networks
are asynchronous and unquantised. Furthermore, digital multipliers occupy a large
silicon area, giving a low synapse count on a single chip.
The advantages of analog circuitry are that asynchronous behaviour and smooth
neural activation are automatic. Circuit elements can be small, but noise immunity
is relatively low and arbitrarily high precision is not possible. Most importantly, no
reliable analog, non - volatile memory technology is as yet readily available. For
this reason, learning networks lend themselves more naturally to digital design and
implementation.
Several groups are developing neural chips and boards, and the following listing
does not pretend to be exhaustive. It is included, rather, to indicate the spread of
activity in this field. Analog techniques have been used to build resistor I operational amplifier networks [2,3] similar to those proposed by Hopfield and Tank [4].
A large group at Caltech is developing networks implementing early vision and
auditory processing functions using the intrinsic nonlinearities of MaS transistors in
the subthreshold regime [5,6]. The problem of implementing analog networks with
electrically programmable synapses has been addressed using CCDIMNOS technology [7]. Finally, Garth [8] is developing a digital neural accelerator board ("Netsim") that is effectively a fast SIMD processor with supporting memory and communications chips.
Bit - serial vs. bit - parallel: Bit - serial arithmetic and communication is efficient
for computational processes, allowing good communication within and between
VLSI chips and tightly pipelined arithmetic structures. It is ideal for neural networks as it minimises the interconnect requirement by eliminating multi - wire
busses. Although a bit - parallel design would be free from computational latency
(delay between input and output), pipelining makes optimal use of the high bit rates possible in serial systems, and makes for efficient circuit usage.
2.1 An asynchronous pulse stream VLSI neural network:
In addition to the digital system that forms the substance of this paper, we are
developing a hybrid analOg/digital network family. This work is outlined here, and
has been reported in greater detail elsewhere [9, 10, 11]. The generic (logical and
layout) architecture of a single network of n totally interconnected neurons is shown
575
schematically in figure 1. Neurons are represented by circles, which signal their
states, Vi upward into a matrix of synaptic operators. The state signals are connected to a n - bit horizontal bus running through the synaptic array, with a connection to each synaptic operator in every column. All columns have n operators
(denoted by squares) and each operator adds its synaptic contribution, Tij V j , to the
running total of activity for the neuron i at the foot of the column. The synaptic
function is therefore to multiply the signalling neuron state, Vj , by the synaptic
weight, T ij , and to add this product to the running total. This architecture is common to both the bit - serial and pulse - stream networks.
Synapse
States { Vj }
Neurons
Figure 1. Generic architecture for a network of n totally interconnected neurons.
This type of architecture has many attractions for implementation in 2 - dimensional
j=II -1
silicon as the summation
2
Tij
Vj is distributed in space.
The interconnect
j=O
requirement (n inputs to each neuron) is therefore distributed through a column,
reducing the need for long - range wiring. The architecture is modular, regular and
can be easily expanded.
In the hybrid analog/digital system, the circuitry uses a "pulse stream" signalling
method similar to that in a natural neural system. Neurons indicate their state by
the presence or absence of pulses on their outputs, and synaptic weighting is
achieved by time - chopping the presynaptic pulse stream prior to adding it to the
postsynaptic activity summation. It is therefore asynchronous and imposes no fundamental limitations on the activation or neural state. Figure 2 shows the pulse
stream mechanism in more detail. The synaptic weight is stored in digital memory
local to the operator. Each synaptic operator has an excitatory and inhibitory pulse
stream input and output. The resultant product of a synaptic operation, Tij V j , is
added to the running total propagating down either the excitatory or inhibitory
channel. One binary bit (the MSBit) of the stored Tij determines whether the contribution is excitatory or inhibitory.
The incoming excitatory and inhibitory pulse stream inputs to a neuron are
integrated to give a neural activation potential that varies smoothly from 0 to 5 V.
This potential controls a feedback loop with an odd number of logic inversions and
576
.
??
XT ??
V,
.u.u,
?
Figure 2. Pulse stream arithmetic. Neurons are denoted by 0 and synaptic operators
by D.
thus forms a switched "ring - oscillator". H the inhibitory input dominates, the feedback loop is broken. H excitatory spikes subsequently dominate at the input, the
neural activity rises to 5V and the feedback loop oscillates with a period determined
by a delay around the loop. The resultant periodic waveform is then converted to a
series of voltage spikes, whose pulse rate represents the neural state, Vi' Interestingly, a not dissimilar technique is reported elsewhere in this volume, although the
synapse function is executed differently [12].
3. A 5 - STATE BIT - SERIAL NEURAL NETWORK
The overall architecture of the 5 - state bit - serial neural network is identical to
that of the pulse stream network. It is an array of n 2 interconnected synchronous
synaptic operators, and whereas the pulse stream method allowed Vj to assume all
values between "off' and "on", the 5 - state network VJ is constrained to 0, ?0.5 Qr
? 1. The resultant activation function is shown in Figure 3. Full digital multiplication is costly in silicon area, but multiplication of Tij by Vj = 0.5 merely requires
the synaptic weight to be right - shifted by 1 bit. Similarly, multiplication by 0.25
involves a further right - shift of Til' and multiplication by 0.0 is trivially easy. VJ
< 0 is not problematic, as a switchable adder/subtractor is not much more complex
than an adder. Five neural states are therefore feasible with circuitry that is only
slightly more complex than a simple serial adder. The neural state expands from a 1
bit to a 3 bit (5 - state) representation, where the bits represent "add/subtract?",
"shift?" and "multiply by O?".
Figure 4 shows part of the synaptic array. Each synaptic operator includes an 8 bit
shift register memory block holding the synaptic weight, Til' A 3 bit bus for the 5
neural states runs horizontally above each synaptic row. Single phase dynamic
CMOS has been used with a clock frequency in excess of 20 MHz [13). Details of
a synaptic operator are shown in figure 5. The synaptic weight Til cycles around the
shift register and the neural state Vj is present on the state bus. During the first
clock CYCle, the synaptic weight is multiplied by the neural state and during the
second, the most significant bit (MSBit) of the resultant Tij Vj is sign - extended for
577
State VJ
lHRESHOLD
.....-------=-------..
s?
"5
Activity sJ
STATE"
"Sharper"
"Smoother"
~.....::~-"'--x.&..t------
Activity "J
Figure 3. "Hard - threshold", 5 - state and sigmoid activation functions.
J-a-1 T v
~
J-li
.. J
v,
v,
Figure 4. Section of the synaptic array of the 5 - state activation function neural network.
8 bits to allow for word growth in the running summation. A least significant bit
(LSBit) signal running down the synaptic columns indicates the arrival of the LSBit
of the Xj running total. If the neural state is ?O.5 the synaptic weight is right
shifted by 1 bit and then added to or subtracted from the running total. A multiplication of ? 1 adds or subtracts the weight from the total and multiplication by 0
578
.0.5
.0.0
Add/Subtract
Add!
Subtract
Carry
Figure S. The synaptic operator with a 5 - state activation function.
does not alter the running summation.
The final summation at the foot of the column is thresholded externally according
to the 5 - state activation function in figure 3. As the neuron activity Xj' increases
through a threshold value x" ideal sigmoidal activation represents a smooth switch
of neural state from -1 to 1. The 5 - state "staircase" function gives a superficially
much better approximation to the sigmoid form than a (much simpler to implement) threshold function. The sharpness of the transition can be controlled to
"tune" the neural dynamics for learning and computation. The control parameter is
referred to as temperature by analogy with statistical functions with this sigmoidal
form. High "temperature" gives a smoother staircase and sigmoid, while a temperature of 0 reduces both to the ''Hopfield'' - like threshold function. The effects of
temperature on both learning and recall for the threshold and 5 - state activation
options are discussed in section 4.
4. LEARNING AND RECALL WITH VLSI CONSTRAINTS
Before implementing the reduced - arithmetic network in VLSI, simulation experiments were conducted to verify that the 5 - state model represented a worthwhile
enhancement over simple threshold activation. The "benchmark" problem was
chosen for its ubiquitousness, rather than for its intrinsic value. The implications
for learning and recall of the 5 - state model, the threshold (2 - state) model and
smooth sigmoidal activation ( 00 - state) were compared at varying temperatures
with a restricted dynamic range for the weights T ij ? In each simulation a totally
interconnected 64 node network attempted to learn 32 random patterns using the
delta rule learning algorithm (see for example [14]). Each pattern was then corrupted with 25% noise and recall attempted to probe the content addressable
memory properties under the three different activation options.
During learning, individual weights can become large (positive or negative). When
weights are "driven" beyond the maximum value in a hardware implementation,
579
which is determined by the size of the synaptic weight blocks, some limiting
mechanism must be introduced. For example, with eight bit weight registers, the
limitation is -128 S Tij S 127. With integer weights, this can be seen to be a problem of dynamic range, where it is the relationship between the smallest possible
weight (? 1) and the largest (+ 127/-128) that is the issue.
Results: Fig. 6 shows examples of the results obtained, studying learning using 5 state activation at different temperatures, and recall using both 5 - state and threshold activation. At temperature T=O, the 5 - state and threshold models are
degenerate, and the results identical. Increasing smoothness of activation (temperature) during learning improves the quality of learning regardless of the activation
function used in recall, as more patterns are recognised successfully. Using 5 - state
activation in recall is more effective than simple threshold activation. The effect of
dynamic range restrictions can be assessed from the horizontal axis, where T/j:6. is
shown. The results from these and many other experiments may be summarised as
follows:5 - State activation vs. threshold:
1) Learning with 5 - state activation was protracted over the threshold activation,
as binary patterns were being learnt, and the inclusion of intermediate values
added extra degrees of freedom.
2) Weight sets learnt using the 5 - state activation function were "better" than
those learnt via threshold activation, as the recall properties of both 5 - state
and threshold networks using such a weight set were more robust against
noise.
3) Full sigmoidal activation was better than 5 - state, but the enhancement was
less significant than that incurred by moving from threshold - 5 - state. This
suggests that the law of diminishing returns applies to addition of levels to the
neural state Vi' This issue has been studied mathematically [15], with results
that agree qualitatively with ours.
Weight Saturation:
Three methods were tried to deal with weight saturation. Firstly, inclusion of a
decay, or "forgetting" term was included in the learning cycle [1]. It is our view
that this technique can produce the desired weight limiting property, but in the time
available for experiments, we were unable to "tune" the rate of decay sufficiently
well to confirm it. Renormalisation of the weights (division to bring large weights
back into the dynamic range) was very unsuccessful, suggesting that information
distributed throughout the numerically small weights was being destroyed. Finally,
the weights were allowed to "clip" (ie any weight outside the dynamic range was set
to the maximum allowed value). This method proved very successful, as the learning algorithm adjusted the weights over which it still had control to compensate for
the saturation effect. It is interesting to note that other experiments have indicated
that Hopfield nets can "forget" in a different way, under different learning control,
giving preference to recently acquired memories [16]. The results from the saturation experiments were:1) For the 32 pattemJ64 node problem, integer weights with a dynamic range
greater than ?30 were necessary to give enough storage capability.
2) For weights with maximum values TiJ = 50-70, "clipping" occurs, but network performance is not seriously degraded over that with an unrestricted
weight set.
580
15
c 10
"0
~
e
,
=
.2
-
~
T=20
T=10
T=O
....--- -- - -
- - ,- ., ...
;A .......;.. f:'-:' :::::7.:::.::-:::-: f'-.
I
en
e
u
T=30 _._.-.-
15
,...?.-.....-.?. _.?. -.-._.-..
,. ".'
i
,..
i
j''''-,,'i
~-------------
!
!
,
5
-..,..
i
j
'" ?????? ? ???????????????? ??????
j
I
I
,
I
0
O~~~~--~~
0
20
30
40 50 60 70
Limit
5 . state activation function recal1
o
__~~__
40 50 60 70
Limit
tlHopficld" activation function recall
20
30
Figure 6. Recall of patterns learned with the 5 . state activation function and subsequently restored using the 5-state and the hard - threshold activation functions.
T is the "temperature", or smoothness of the activation function, and "limit" the value
ofTI; ?
These results showed that the 5 - state model was worthy of implementation as a
VLSI neural board, and suggested that 8 - bit weights were sufficient.
S. PROJECTED SPECIFICATION OF A HARDWARE NEURAL BOARD
The specification of a 64 neuron board is given here, using a 5 - state bit - serial 64
x 64 synapse array with a derated clock speed of 20 MHz. The synaptic weights are
8 bit words and the word length of the running summation XI is 16 bits to allow for
growth. A 64 synapse column has a computational latency of 80 clock cycles or
bits, giving an update time of 4 .... s for the network. The time to load the weights
into the array is limited to 6O .... s by the supporting RAM, with an access time of
12Ons. These load and update times mean that the network is executing 1 x 10'
operations/second, where one operation is ? Tlj Vj ? This is much faster than a
natural neural network, and much faster than is necessary in a hardware accelerator. We have therefore developed a "paging" architecture, that effectively "trades off" some of this excessive speed against increased network size.
A "moving - patch" neural board: An array of the 5 - state synapses is currently
being fabricated as a VLSI integrated circuit. The shift registers and the
adderlsubtractor for each synapse occupy a disappointingly large silicon area, allowing only a 3 x 9 synaptic array. To achieve a suitable size neural network from this
array, several chips need to be included on a board with memory and control circuitry. The "moving patch" concept is shown in figure 7, where a small array of
synapses is passed over a much larger n x n synaptic array.
Each time the array is "moved" to represent another set of synapses, new weights
must be loaded into it. For example, the first set of weights will be T 11 ?. , T;J ... T 21
... T 2j to Tjj , the second set T j + 1 ,l to Tu etc.. The final weight to be loaded will be
581
Smaller "Patch"
n neurons .. om synaptic array
moves over array
rr~
~'-
_____)__-..
>
Figure 7. The "moving patch" concept, passing a small synaptic "patch" over a larger
run synapse array.
TNt? Static, off - the - shelf RAM is used to store the weights and the whole opera-
tion is pipelined for maximum efficiency. Figure 8 shows the board level design for
the network.
Control
Synaptic Accelerator Chips
HOST
Figure 8. A "moving patch" neural network board.
The small "patch" that moves around the array to give n neurons comprises 4 VLSI
synaptic accelerator chips to give a 6 x 18 synaptic array. The number of neurons to
be simulated is 256 and the weights for these are stored in 0.5 Mb of RAM with a
load time of 8ms. For each "patch" movement, the partial runnin~ summatinn ;.
582
calculated for each column, is stored in a separate RAM until it is required to be
added into the next appropriate summation. The update time for the board is 3ms
giving 2 x 107 operations/second. This is slower than the 64 neuron specification,
but the network is 16 times larger, as the arithmetic elements are being used more
efficiently. To achieve a network of greater than 256 neurons, more RAM is
required to store the weights. The network is then slower unless a larger number of
accelerator chips is used to give a larger moving "patch".
6. CONCLUSIONS
A strategy and design method has been given for the construction of bit - serial
VLSI neural network chips and circuit boards. Bit - serial arithmetic, coupled to a
reduced arithmetic style, enhances the level of integration possible beyond more
conventional digital, bit - parallel schemes. The restrictions imposed on both synaptic weight size and arithmetic precision by VLSI constraints have been examined
and shown to be tolerable, using the associative memory problem as a test.
While we believe our digital approach to represent a good compromise between
arithmetic accuracy and circuit complexity, we acknowledge that the level of
integration is disappointingly low. It is our belief that, while digital approaches
may be interesting and useful in the medium term, essentially as hardware accelerators for neural simulations, analog techniques represent the best ultimate option in 2
- dimensional silicon. To this end, we are currently pursuing techniques for analog
pseudo - static memory, using standard CMOS technology. In any event, the full
development of a nonvolatile analog memory technology, such as the MNOS technique [7], is key to the long - term future of VLSI neural nets that can learn.
7. ACKNOWLEDGEMENTS
The authors acknowledge the support of the Science and Engineering Research
Council (UK) in the execution of this work.
References
1.
2.
3.
4.
5.
6.
S. Grossberg, "Some Physiological and Biochemical Consequences of Psychological Postulates," Proc. Natl. Acad. Sci. USA, vol. 60, pp. 758 - 765, 1968.
H. P. Graf, L. D. Jackel, R. E. Howard, B. Straughn, J. S. Denker, W.
Hubbard, D. M. Tennant, and D. Schwartz, "VLSI Implementation of a
Neural Network Memory with Several Hundreds of Neurons," Proc. AlP
Conference on Neural Networks for Computing. Snowbird, pp. 182 - 187, 1986.
W. S. Mackie, H. P. Graf, and J. S. Denker, "Microelectronic Implementation of Connectionist Neural Network Models," IEEE Conference on Neural
Information Processing Systems. Denver, 1987.
J . J. Hopfield and D. W. Tank, "Neural" Computation of Decisions in Optimisation Problems," BioI. Cybern., vol. 52, pp. 141 - 152, 1985.
M. A. Sivilotti, M. A. Mahowald, and C. A. Mead, Real - Time Visual Computations Using Analog CMOS Processing Arrays, 1987. To be published
C. A. Mead, "Networks for Real - Time Sensory Processing," IEEE Conference on Neural Information Processing Systems, Denver, 1987.
583
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
J. P. Sage, K. Thompson. and R. S. Withers, "An Artificial Neural Network
Integrated Circuit Based on MNOSlCCD Principles," Proc. AlP Conference on
Neural Networlcs for Computing, Snowbird, pp. 381 - 385, 1986.
S. C. J. Garth, "A Chipset for High Speed Simulation of Neural Network Systems," IEEE Conference on Neural Networlc.s, San Diego, 1987.
A. F. Murray and A. V. W. Smith, "A Novel Computational and Signalling
Method for VLSI Neural Networks," European Solid State Circuits Conference
, 1987.
A. F. Murray and A. J. W. Smith, "Asynchronous Arithmetic for VLSI
Neural Systems," Electronics Letters, vol. 23, no. 12, p. 642, June, 1987.
A. F. Murray and A. V. W. Smith, "Asynchronous VLSI Neural Networks
using Pulse Stream Arithmetic," IEEE Journal of Solid-State Circuits and Systems, 1988. To be published
M. E. Gaspar, "Pulsed Neural Networks: Hardware, Software and the Hopfield AID Converter Example," IEEE Conference on Neural Information Processing Systems. Denver, 1987.
M. S. McGregor, P. B. Denyer, and A. F. Murray, "A Single - Phase Clocking Scheme for CMOS VLSI," Advanced Research in VLSI " Proceedings of the
1987 Stanford Conference, 1987.
D. E. Rumelhart, G. E. Hinton, and R. J. Williams, "Learning Internal
Representations by Error Propagation," Parallel Distributed Processing "
Explorations in the Microstructure of Cognition, vol. 1, pp. 318 - 362, 1986.
M. Fleisher and E. Levin, "The Hopfiled Model with Multilevel Neurons
Models," IEEE Conference on Neural Information Processing Systems. Denver,
1987.
G. Parisi, "A Memory that Forgets," J. Phys. A .' Math. Gen., vol. 19, pp.
L617 - L620, 1986.
| 27 |@word manageable:1 eliminating:1 inversion:1 chopping:1 pulse:13 simulation:4 tried:1 solid:2 disappointingly:2 carry:1 electronics:1 initial:1 series:1 seriously:1 ours:1 interestingly:1 activation:34 yet:1 must:3 readily:1 update:5 v:3 signalling:3 scotland:1 smith:4 math:1 node:2 preference:1 sigmoidal:4 simpler:1 firstly:1 five:1 along:2 become:1 mayfield:1 acquired:1 forgetting:1 themselves:1 multi:1 totally:4 increasing:1 circuit:9 medium:1 sivilotti:1 developed:1 extremum:1 fabricated:1 pseudo:1 every:1 expands:1 growth:2 oscillates:1 uk:1 control:6 schwartz:1 appear:1 before:1 positive:1 engineering:3 understood:2 multiplex:1 local:1 limit:3 consequence:1 acad:1 mead:2 approximately:1 studied:1 examined:1 suggests:1 limited:1 range:7 grossberg:1 block:2 implement:1 addressable:1 area:4 word:3 road:1 regular:1 pipelined:2 operator:11 storage:1 cybern:1 restriction:2 conventional:1 imposed:1 layout:2 straightforward:2 regardless:1 williams:1 thompson:1 sharpness:1 rule:1 attraction:1 array:22 importantly:1 dominate:1 limiting:2 construction:1 diego:1 us:1 designing:1 element:2 rumelhart:1 electrical:1 fleisher:1 connected:1 cycle:4 trade:1 movement:1 broken:1 complexity:2 microelectronic:1 dynamic:8 compromise:1 division:1 efficiency:1 easily:2 hopfield:5 chip:9 differently:1 represented:2 surrounding:1 fast:1 effective:1 artificial:1 outside:1 exhaustive:1 whose:1 modular:1 larger:5 stanford:1 say:1 ability:2 final:2 associative:1 advantage:3 rr:1 transistor:1 net:2 parisi:1 interconnected:5 product:2 mb:1 tu:1 loop:4 gen:1 degenerate:1 achieve:2 moved:1 gentle:1 qr:1 enhancement:2 requirement:2 produce:1 cmos:4 ring:1 executing:1 mackie:1 propagating:1 minimises:1 snowbird:2 ij:2 odd:1 involves:1 indicate:2 foot:2 waveform:1 functionality:1 subsequently:2 pipelining:1 exploration:1 alp:2 implementing:4 behaviour:1 multilevel:1 microstructure:1 summation:8 mathematically:1 adjusted:1 around:3 sufficiently:1 cognition:1 circuitry:4 early:1 smallest:1 proc:3 currently:2 jackel:1 council:1 hubbard:1 largest:1 successfully:1 rather:2 shelf:1 varying:1 voltage:1 june:1 indicates:1 interconnect:3 biochemical:1 integrated:3 gaspar:1 diminishing:1 vlsi:21 tank:2 issue:3 upward:1 overall:1 denoted:2 development:2 constrained:1 subtractor:1 integration:2 field:1 simd:1 identical:2 represents:4 excessive:1 alter:1 mimic:1 future:1 connectionist:1 fundamentally:1 simultaneously:1 tightly:1 individual:1 phase:2 amplifier:1 freedom:1 clocking:1 multiply:2 natl:1 implication:1 partial:1 necessary:2 unless:1 circle:1 desired:1 psychological:1 increased:1 column:8 mhz:3 disadvantage:1 clipping:1 paging:2 mahowald:1 hundred:1 delay:2 successful:1 levin:1 conducted:1 reported:3 stored:4 varies:1 corrupted:2 periodic:1 learnt:3 synthetic:1 fundamental:1 ie:1 off:4 receiving:1 physic:1 transmitting:1 postulate:1 american:1 til:3 return:1 style:1 li:1 mnos:1 suggesting:1 potential:2 nonlinearities:1 converted:1 includes:1 register:4 vi:5 stream:11 tion:1 view:1 option:3 capability:2 parallel:5 complicated:1 contribution:2 om:1 square:1 opera:1 degraded:1 accuracy:1 loaded:2 efficiently:1 listing:1 subthreshold:1 yield:1 published:2 processor:1 synapsis:7 phys:1 synaptic:37 against:2 frequency:1 pp:6 naturally:1 resultant:4 static:2 auditory:1 proved:1 logical:1 recall:11 knowledge:1 improves:1 back:1 response:1 synapse:11 furthermore:1 clock:4 until:1 horizontal:2 adder:3 propagation:1 quality:1 indicated:1 believe:1 usage:1 effect:3 building:1 requiring:1 multiplier:1 functioning:1 concept:2 staircase:2 verify:1 usa:1 deal:1 attractive:1 wiring:1 during:4 excitation:1 m:3 recognised:1 temperature:9 bring:1 novel:1 recently:1 volatile:1 common:1 sigmoid:3 denver:4 volume:1 analog:14 discussed:2 numerically:1 silicon:9 significant:3 smoothness:2 automatic:1 outlined:2 trivially:1 similarly:1 inclusion:2 language:1 had:1 moving:6 specification:4 stable:1 access:1 operating:1 etc:1 add:6 showed:1 pulsed:1 driven:1 store:2 binary:2 arbitrarily:1 fault:3 caltech:1 seen:1 greater:3 unrestricted:1 period:1 signal:4 arithmetic:12 ii:2 full:3 smoother:2 reduces:1 alan:1 smooth:3 xf:1 faster:2 calculation:1 long:2 compensate:1 host:1 serial:16 netsim:1 controlled:1 ensuring:1 vision:1 essentially:1 optimisation:1 aspire:1 represent:4 achieved:1 addition:2 schematically:1 whereas:1 addressed:1 extra:1 integer:2 presence:1 ideal:2 intermediate:2 enough:2 easy:1 reacts:1 switch:2 automated:1 xj:2 destroyed:1 architecture:9 converter:1 imperfect:1 shift:5 synchronous:3 whether:1 ultimate:1 passed:1 passing:1 programmable:2 useful:2 tij:9 latency:2 listed:1 tune:2 hardware:5 clip:1 reduced:2 occupy:2 problematic:1 inhibitory:6 shifted:2 sign:1 delta:1 summarised:1 vol:5 wafer:1 group:2 key:1 threshold:16 thresholded:1 ram:5 merely:1 quantised:1 run:2 letter:1 family:1 throughout:1 pursuing:1 patch:9 withers:1 decision:1 bit:35 networlc:1 activity:11 strength:1 constraint:2 straughn:1 software:1 networlcs:1 speed:4 expanded:1 relatively:2 department:1 developing:4 according:1 electrically:1 smaller:1 slightly:1 postsynaptic:1 restricted:1 recourse:1 agree:1 bus:4 count:1 mechanism:2 merit:1 end:1 unusual:1 studying:1 junction:1 available:2 operation:4 multiplied:1 probe:1 eight:1 worthwhile:1 denker:2 generic:2 appropriate:1 tolerable:1 tlj:1 subtracted:1 slower:2 assumes:1 running:10 pretend:1 giving:5 murray:5 build:2 move:2 added:4 zoe:1 spike:2 occurs:1 primary:1 costly:1 restored:1 strategy:1 enhances:1 unable:1 separate:1 simulated:1 sci:1 presynaptic:1 reason:1 length:1 relationship:1 executed:1 demerit:1 sharper:1 holding:1 negative:1 rise:2 sage:1 design:8 implementation:7 allowing:2 neuron:27 wire:1 benchmark:1 acknowledge:2 howard:1 supporting:2 extended:2 incorporated:1 communication:3 hinton:1 worthy:1 tnt:1 introduced:1 cast:1 required:2 immunity:2 connection:1 learned:2 chipset:1 able:1 beyond:2 suggested:1 parallelism:1 below:1 pattern:5 regime:1 challenge:1 saturation:4 reliable:1 memory:13 lend:1 unsuccessful:1 belief:1 suitable:1 event:1 natural:4 hybrid:3 protracted:1 advanced:2 scheme:2 interconnects:1 technology:5 axis:1 coupled:1 prior:1 acknowledgement:1 multiplication:7 graf:2 law:1 accelerator:6 limitation:2 proportional:1 interesting:2 analogy:1 tjj:1 digital:20 switched:1 incurred:1 degree:1 sufficient:1 imposes:1 switchable:1 principle:1 squashing:1 row:1 excitatory:6 elsewhere:2 asynchronous:6 free:1 allow:2 institute:1 edinburgh:2 tolerance:2 distributed:5 feedback:3 calculated:1 superficially:1 transition:1 sensory:1 author:1 qualitatively:1 projected:2 san:1 subtracts:1 garth:2 excess:1 sj:1 compact:1 ons:1 logic:1 confirm:1 incoming:1 xi:5 butler:1 stimulated:1 learn:3 channel:1 robust:1 operational:1 complex:2 european:1 anthony:1 vj:11 spread:1 whole:1 noise:4 arrival:1 repeated:1 allowed:3 fig:1 referred:1 en:1 board:13 nonvolatile:1 aid:1 precision:2 comprises:1 resistor:1 forgets:1 weighting:2 externally:1 down:2 load:3 xt:1 denyer:1 substance:1 decay:2 physiological:1 unattractive:1 dominates:1 intrinsic:2 adding:1 effectively:2 tennant:1 execution:1 subtract:3 smoothly:2 forget:1 visual:1 horizontally:1 applies:1 determines:2 ma:1 bioi:1 king:1 oscillator:1 absence:1 feasible:1 change:2 hard:2 included:3 determined:2 content:1 reducing:1 degradation:2 total:7 attempted:2 internal:1 support:1 arises:1 assessed:1 dissimilar:1 mcgregor:1 |
1,872 | 270 | 660
Geiger and Girosi
Coupled Markov Random Fields and
Mean Field Theory
Davi Geigerl
Artificial Intelligence
Laboratory, MIT
545 Tech. Sq. # 792
Cambridge, MA 02139
and
Federico Girosi
Artificial Intelligence
Laboratory, MIT
545 Tech. Sq. # 788
Cambridge, MA 02139
ABSTRACT
In recent years many researchers have investigated the use of Markov
Random Fields (MRFs) for computer vision. They can be applied
for example to reconstruct surfaces from sparse and noisy depth
data coming from the output of a visual process, or to integrate
early vision processes to label physical discontinuities. In this paper we show that by applying mean field theory to those MRFs
models a class of neural networks is obtained. Those networks can
speed up the solution for the MRFs models. The method is not
restricted to computer vision.
1
Introduction
In recent years many researchers (Geman and Geman, 1984) (Marroquin et. al.
1987) (Gamble et. al. 1989) have investigated the use of Markov Random Fields
(MRFs) for early vision. Coupled MRFs models can be used for the reconstruction
of a function starting from a set of noisy sparse data, such as intensity, stereo, or
motion data. They have also been used to integrate early vision processes to label
physical discontinuities. Two fields are usually required in the MRFs formulation
of a problem: one represents the function that has to be reconstructed, and the
other is associated to its discontinuities. The reconstructed function, say I, has
1 New
address is Siemens Corporate Research, 755 College Road East, Princeton NJ 08540
Coupled Markov Random Fields and Mean Field Theory
Figure 1: The 6quare lattice with the line proceu I and the field
pizel6.
J defined at 60me
a continuous range and the discontinuity field, say I, is a binary field (1 if there
is a discontinuity and 0 otherwise, see figure 1). The essence of the MRFs model
is that the probability distribution of the configuration of the fields, for a given
a set of data, has a Gibbs distribution for some cost functional dependent upon
a small neighborhood. Since the fields have a discrete range, to find the solution
becomes a combinatorial optimization problem, that can be solved by means of
methods like the Monte Carlo one (simulated annealing (Kirkpatrick and all, 1983),
for example). However it has a main drawback: the amount of computer time
needed for the implementation.
We propose to approximate the solution of the problem formulated in the MRFs
frame with its "average solution." The mean field theory (MFT) allows us to find
deterministic equations for MRFs whose solution approximates the solution of the
statistical problem. A class of neural networks can naturally solve these equations
(Hopfield, 1984) (Koch et. al., 1985) (Geiger and Yuille, 1989). An advantage of
such an approach is that the solution of the networks is faster than the Monte Carlo
techniques, commonly used to deal with MRFs.
A main novelty in this work, and a quite general one, is to show that the binary
field representing the discontinuities can be averaged out to yield an effective theory independent of the binary field. The possibility of writing a set of equations
describing the network is also useful for a better understanding of the nature of the
solution and of the parameters of the model. We show the network performance in
an example of image reconstruction from sparse data.
661
662
Geiger and Girosi
2
MRFs and Bayes approach
One of the main attractions of MRFs models in vision is that they can deal directly
with discontinuities. We consider coupled MRFs depending upon two fields, J
and I. For the problem of image reconstruction the field J represents the field to
be smoothed and I represents the discontinuities. In this case I is a binary field,
assuming the values 1 if there is a discontinuity and 0 otherwise. The Markov
property asserts that the probability of a certain value of the field at any given
site in the lattice depends only upon neighboring sites. According to the CliffordHammersley theorem, the prior probability of a state of the fields J and I has the
Gibbs form:
P(j, I)
1
= _e-fjU(J,I)
Z
(2.1)
where J and I are the fields, e.g. the surface-field and its discontinuities, Z is the
normalization constant also known as the partition function, U(j, I) = Ei Ui(J, I)
is an energy function that can be computed as the sum of local contributions from
each lattice site i, and f3 is a parameter that is called the inverse of the natural
temperature of the field. If a sparse observation 9 for any given surface-field / is
given and a model of the noise is available then one knows the conditional probability
P(gIJ, I). Bayes theorem then allows us to write the posterior distribution:
.!.
-fjv(JI,)
P(J II ) = P(gIJ, I)P(j, I) =
, 9
P(g)
- Ze
.
(2.2)
For the case of a sparse image corrupted by white gaussian noise
V(j,llg) = L~i(ji _gi)2
+ Ui(j,l)
(2.3)
i
=
where ~i;
1 or 0 depending on whether data are available or not. V(J,llg) is
sometimes called the visual cost !unction. The solution for the problem is the given
by some estimate of the fields. The maximum of the posterior distribution or other
related estimates of the "true" data-field value can not be computed analytically,
but sample distributions of the field with the probability distribution of (2.2) can
be obtained using Monte Carlo techniques such as the Metropolis algorithm. These
algorithms sample the space of possible values of the fields accwding to the proba.bility distribution P(j,llg).
A drawback of coupled MRFs has been the amount of computer time used in the
Metropolis algorithm or in simulated annea.ling (Kirkpatrick et. al., 1983).
A justification for using the mean field (MF) as a measure of the fields, J for example, resides in the fact that it represents the minimum variance Bayes estimator.
More precisely, the average variance of the field J is given by
Coupled Markov Random Fields and Mean Field Theory
Va".!
= LU -
/)2 PU, llg)
I,l
where / is a given estimate of the field, the EJ,l represents the sum over all the
possible configurations of / and " and Va".! is the variance. Minimizing Va".! with
respect to all possible values of / we obtain
This equation for / defines the deterministic MF equations.
2.1
MFT and Neural Networks
To connect MRFs to neural networks, we use Mean field theory (MFT) to obtain
deterministic equations from MRFs that represent a class of neural networks.
The mean field for the values f and I at site i are given by
" = L "PU, llg)
and
r. = L
1.1
The sum over the binary process, Ii
approximation,
Ii =
L
e-tn~i(J'-"i)2+Ui(J.f#i.I,=1)]
----Z-.- - -
(2.5)
?
where the partition function Z where factorized as
=L
(2.4)
0,1 gives for (2.3), using the mean field
I
Zi
'i P(/, Ilg)
e- fJ >'i(Ji-"i)2 (e-fJUi(/,f#,.li=O)
TIi Zi'
In this case
+ e- fJUi (J,T#i,I,=l?).
I
Another way to write the equation for / is
_fJV.-llecti ..
"
" = L.J"
I
where
e
?
Z.
?
(2.6)
663
664
Geiger and Girosi
The important result obtained here is that the effective potential does not dependend on the binary field Ii. The line process field has been eliminated to yield a
temperature dependent effective potential (also called visual cost function). The
interaction of the field f with itself has changed after the line process has been
averaged out. We interpret this result as the effect of the interaction of the line
processes with the field f to yield a new temperature dependent potential.
The computation of the sum over all the configurations of the field f is hard and
we use the saddle point approximation. In this case is equivalent to minimize
veJJeeti""(f). A dynamical equation to find the minimum of veJJeeti'Oe is given by
introducing a damping force
that brings the system to equilibrium. Therefore the
mean field equation under the mean field and saddle point approximation becomes
*
.!!.... ~eJJeeti'Oe(1 r = 8h
8h ?
,'J
(2.8)
8t
Equation (2.8) represents a class of unsupervised neural networks coupled to (2.5).
The mean field solution is given by the fixed point of (2.8) and (2.5) it is attained
after running (2.8) and (2.5) as t ........ 00. This network is better understood with an
example of image reconstruction.
3
Example: Image reconstruction
To reconstruct images from sparse data and to detect discontinuities we use the
weak membrane model where Ui(J, I) in two dimensions is given by
u..'" ?(f, h , v)
=
Q
~[(-I. . -
L...J J."
i,;
J',,. . 1)2(1-h.'" ')+(/''" '- J.-"
-I. 1 .)2(1_v . .)]+"V(J.. ?+V? .)
'"
'''i"
'"
I
(3.1)
and
Q
and l' are positive parameters.
The first term, contains the interaction between the field and the line processes: if
the horizon tal or vertical gradient is very high at site (i, j) the corresponding line
1), to make the visual cost
process will be very likely to be active (~,; 1 or Vi,;
function decrease and signal a discontinuity. The second term takes into account
the price we pay each time we create a discontinuity and is necessary to prevent
the creation of discontinuities everywhere. The effective cost function (2.7) then
becomes
=
=
Coupled Markov Random Fields and Mean Field Theory
665
Figure 2: The network i& repre,ented for the one dimen&ional ca,e. The line, are
the connection,
Vai" = ~ ["\ii(Ji,i-9i,i )2+a(a~i)2+(ai,i)2- ~ln[(I+e-"('Y-a4t/?)(1+e-"('Y-a4i./?]]
',J
(3.2)
where a~i
-
= Ii.; -
fi-1,i, ar,i
= Ii,i -
1
----..",.....~-',J - 1 + e"('Y-a(f'.j-la-l,j)2)
h? . -
Ji';-l
and
and (2.5) is then given by
1
, = 1 + e"('Y- a ( J,'.j- f,'.j-d 2 )
Vi i
(3.3).
we point out here that while the line process field is a binary field, its mean value
is a continuous (analog) function in the range between 0 and 1.
Discretizing (2.8) in time and applying for (3.2), we obtain
1.';+1 =
I.j - w [..\ii(h~i -
-a(l.~i -1."_1,;)(1 where h.,i
algorithm.
by (3.4) at
J again at
a(l.~i -1.~i-1)(1 - v~i) + a(l.~i+1 -1.~i)(1 - v~i+d
hf.i) + a(l."+1'i -1.~i)(l- hf+1,i)]
9i,i) -
and vi,i are given by the network (3.3) and n is the time step on the
We notice that (3.4) is coupled with (3.3) such that the field fis updated
step n and then (3.3) updates the field h and v before (3.4) updates field
step n + 1.
This is a simple unsupervised neural network where the imput are the fields J and
the output is the line process field h or v. This network is coupled to the network
(2.8) to solve for the field J and then constitute the global network for this problem
(see figure 2). It has been shown by many authors and (Geiger and Yuille, 1989)
that these class of networks is equivalent to Hcpfield networks (Hopfield, 1984)
(Koch et. al., 1985).
(3.4)
666
Geiger and Girosi
Figure 3: a. The .dill life image 128 x 128 pizel6. The image 6moothed with
I = 1400 and Q = 4 for 9 iteration6. The line proceS6 field (needs thinning). b. A
face image of 128 x 128 pizel6. Randomly chosen 50 % of the original image (for
di6play the other 50% are filled with white dot6). c. The network described above i6
applied to 6mooth and fill in using the same parameters and for 10 iterations. For
comparison we show the results of simply bluring the 6par6e data (no line process
field).
An important connection we make is to show (Geiger and Girosi, 1989) (Geiger,
1989) that the work of Blake and Zisserman (Blake and Zisserman, 1987) can be
seen as an approximation of these results.
Coupled Markov Random Fields and Mean Field Theory
In the zero temperature limit (f3 -+ 00) (3.3) becomes the Heaviside function (1
or 0) and the interpretation is simple: when the horizontal or vertical gradient are
larger than a threshold (JI) a vertical or horizontal discontinuity is created.
4
Results
We applied the network to a real still life image and the result was an enhancement
of specular edges, shadow edges and some other contours while smoothing out the
noise (see Figure 3a). This result is consistent with all the images we have used.
From one face image we produced sparse data by randomly suppressing 50% of
the data. (see Figure 3b). We then applied the neural network to reconstruct the
lmage.
AcknowledgeIllents
We are grateful to Tomaso Poggio for his guidance and support.
References
A. Blake and A.
MIT Press.
Zisserman. (1987) Vi,mal Recondruction. Cambridge, Mass:
E. Gamble and D. Geiger and T. Poggio and D. Weinshall. (1989) Integration of
vision modulea and labeling of aurface diacontinuitiea. Invited paper to IEEE Trans.
Sustems, Man & Cybernetics, December.
D. Geiger and F. Girosi. (1989) Parallel and deterministic algorithma for MRFs:
surface reconstruction and integration. A.!, Memo No.1114. Artificial Intelligence
Laboratory of MIT.
D. Geiger. (1989) Viaual modela with datiatical field theory. Ph.D. thesis. MIT,
Physics department and Artificial Intelligence Laboratory.
D. Geiger and A. Yuille. (1989) A common framework for image segmentation
and surface recon..truction. Harvard Robotics Laboratory Technical Report, 89-7,
Harvard, August.
S. Geman and D. Geman. (1984) Stochadic Relazation, Gibba Dutributiona, and
the Bayeaian Redoration of Imagea. Pattern Analysis and Machine Intelligence,
PAMI-6:721-741.
J. J. Hopfield. (1984) Neurona with graded reaponse have collective computational
properties like those of two-state neurona. Proc. Natl. Acad. ScL, 81:3088-3092,
S. Kirkpatrick and C. D. Gelatt and M. P.
Simulated Annealina. Science. 220:219-227.
Vecchio
(1983) Optimization by
C. Koch and J. Marroquin and A. Yuille. (1985) Analog 'Neuronal' Networka in
Early Vision. Proc. Natl. Acad. SeL, 83:4263-4267.
J. L. Marroquin and S. Mitter and T. Poggio. (1987) Probabilutic Solution of
nl-Poled Problema in Computational Viaion. J. Amer. Stat. Assoc., 82:76-89.
667
| 270 |@word configuration:3 contains:1 suppressing:1 unction:1 partition:2 girosi:7 update:2 davi:1 intelligence:5 ional:1 dimen:1 tomaso:1 bility:1 becomes:4 factorized:1 mass:1 weinshall:1 nj:1 assoc:1 positive:1 before:1 understood:1 local:1 limit:1 acad:2 pami:1 range:3 averaged:2 sq:2 road:1 applying:2 writing:1 equivalent:2 deterministic:4 neurona:2 starting:1 estimator:1 attraction:1 fill:1 his:1 justification:1 updated:1 harvard:2 ze:1 geman:4 solved:1 mal:1 oe:2 decrease:1 ui:4 grateful:1 creation:1 yuille:4 upon:3 hopfield:3 effective:4 monte:3 artificial:4 labeling:1 neighborhood:1 whose:1 quite:1 larger:1 solve:2 say:2 reconstruct:3 federico:1 otherwise:2 mooth:1 gi:1 dependend:1 itself:1 noisy:2 advantage:1 reconstruction:6 propose:1 interaction:3 coming:1 neighboring:1 asserts:1 enhancement:1 depending:2 stat:1 quare:1 shadow:1 drawback:2 dill:1 koch:3 blake:3 equilibrium:1 early:4 proc:2 label:2 combinatorial:1 ilg:1 create:1 mit:5 gaussian:1 ej:1 sel:1 tech:2 detect:1 fjv:2 mrfs:17 dependent:3 fis:1 smoothing:1 integration:2 field:63 f3:2 eliminated:1 represents:6 unsupervised:2 report:1 randomly:2 proba:1 possibility:1 llg:5 kirkpatrick:3 nl:1 natl:2 edge:2 necessary:1 poggio:3 damping:1 filled:1 guidance:1 ar:1 lattice:3 cost:5 introducing:1 connect:1 corrupted:1 physic:1 vecchio:1 again:1 thesis:1 li:1 account:1 potential:3 tii:1 depends:1 vi:4 repre:1 bayes:3 hf:2 parallel:1 contribution:1 minimize:1 variance:3 yield:3 weak:1 produced:1 lu:1 carlo:3 researcher:2 cybernetics:1 energy:1 naturally:1 associated:1 modela:1 segmentation:1 marroquin:3 thinning:1 attained:1 zisserman:3 formulation:1 amer:1 horizontal:2 ei:1 defines:1 brings:1 effect:1 true:1 analytically:1 laboratory:5 deal:2 white:2 essence:1 tn:1 motion:1 temperature:4 fj:1 image:14 fi:1 common:1 functional:1 physical:2 ji:6 analog:2 interpretation:1 approximates:1 interpret:1 mft:3 cambridge:3 gibbs:2 ai:1 i6:1 surface:5 pu:2 posterior:2 recent:2 certain:1 binary:7 discretizing:1 life:2 seen:1 minimum:2 novelty:1 signal:1 ii:8 corporate:1 technical:1 faster:1 va:3 vision:8 iteration:1 normalization:1 sometimes:1 represent:1 robotics:1 annealing:1 invited:1 december:1 zi:2 specular:1 whether:1 stereo:1 constitute:1 useful:1 amount:2 ph:1 recon:1 notice:1 discrete:1 write:2 threshold:1 prevent:1 year:2 sum:4 inverse:1 everywhere:1 geiger:12 pay:1 precisely:1 tal:1 imput:1 speed:1 department:1 according:1 membrane:1 metropolis:2 restricted:1 ln:1 equation:10 describing:1 needed:1 know:1 scl:1 available:2 gelatt:1 original:1 running:1 graded:1 gradient:2 simulated:3 me:1 assuming:1 minimizing:1 memo:1 vai:1 implementation:1 collective:1 vertical:3 observation:1 relazation:1 markov:8 moothed:1 frame:1 smoothed:1 august:1 intensity:1 required:1 connection:2 discontinuity:15 trans:1 address:1 usually:1 dynamical:1 pattern:1 natural:1 force:1 representing:1 created:1 coupled:11 prior:1 understanding:1 integrate:2 consistent:1 changed:1 face:2 sparse:7 depth:1 dimension:1 resides:1 contour:1 author:1 commonly:1 reconstructed:2 approximate:1 global:1 active:1 continuous:2 nature:1 ca:1 investigated:2 main:3 noise:3 ling:1 neuronal:1 site:5 mitter:1 theorem:2 horizon:1 mf:2 simply:1 saddle:2 likely:1 visual:4 ma:2 conditional:1 formulated:1 price:1 man:1 hard:1 called:3 gij:2 la:1 siemens:1 east:1 gamble:2 college:1 support:1 heaviside:1 princeton:1 |
1,873 | 2,700 | Generalization Error Bounds for Collaborative
Prediction with Low-Rank Matrices
Nathan Srebro
Department of Computer Science
University of Toronto
Toronto, ON, Canada
[email protected]
Noga Alon
School of Mathematical Sciences
Tel Aviv University
Ramat Aviv, Israel
[email protected]
Tommi S. Jaakkola
Computer Science and Artificial Intelligence Laboratory
Massachusetts Institute of Technology
Cambridge, MA, USA
[email protected]
Abstract
We prove generalization error bounds for predicting entries in a partially
observed matrix by fitting the observed entries with a low-rank matrix. In
justifying the analysis approach we take to obtain the bounds, we present
an example of a class of functions of finite pseudodimension such that
the sums of functions from this class have unbounded pseudodimension.
1
Introduction
?Collaborative filtering? refers to the general task of providing users with information on
what items they might like, or dislike, based on their preferences so far and how they relate
to the preferences of other users. This approach contrasts with a more traditional featurebased approach where predictions are made based on features of the items.
For feature-based approaches, we are accustomed to studying prediction methods in terms
of probabilistic post-hoc generalization error bounds. Such results provide us a (probabilistic) bound on the performance of our predictor on future examples, in terms of its
performance on the training data. These bounds hold without any assumptions on the true
?model?, that is the true dependence of the labels on the features, other than the central
assumptions that the training examples are drawn i.i.d. from the distribution of interest.
In this paper we suggest studying the generalization ability of collaborative prediction
methods. By ?collaborative prediction? we indicate that the objective is to be able to predict user preferences for items, that is, entries in some unknown target matrix Y of useritem ?ratings?, based on observing a subset YS of the entries in this matrix1 . We present
1
In other collaborative filtering tasks, the objective is to be able to provide each user with a few
items that overlap his top-rated items, while it is not important to be able to correctly predict the users
ratings for other items. Note that it is possible to derive generalization error bounds for this objective
based on bounds for the ?prediction? objective.
arbitrary source distribution
random training set
hypothesis
training error
generalization error
?
target matrix Y
? random set S of observed entries
?
predicted matrix X
? observed discrepancy DS (X; Y )
?
true discrepancy D(X; Y )
Figure 1: Correspondence with post-hoc bounds on the generalization error for standard
feature-based prediction tasks
Pn Pm
1
bounds on the true average overall error D(X; Y ) = nm
i=1
a=1 loss(Xia ; Yia ) of
the P
predictions X in terms of the average error over the observed entries DS (X; Y ) =
1
ia?S loss(Xia ; Yia ), without making any assumptions on the true nature of the pref|S|
erences Y . What we do assume is that the subset S of entries that we observe is chosen
uniformly at random. This strong assumption parallels the i.i.d. source assumption for
feature-based prediction.
In particular, we present generalization error bounds on prediction using low-rank models.
Collaborative prediction using low-rank models is fairly straight forward. A low-rank matrix X is sought that minimizes the average observed error DS (X; Y ). Unobserved entries
in Y are then predicted according to X. The premise behind such a model is that there
are only a small number of factors influencing the preferences, and that a user?s preference
vector is determined by how each factor applies to that user. Different methods differ in
how they relate real-valued entries in X to preferences in Y , and in the associated measure
of discrepancy. For example, entries in X can be seen as parameters for a probabilistic
models of the entries in Y , either mean parameters [1] or natural parameters [2], and a
maximum likelihood criterion used. Or, other loss functions, such as squared error [3, 2],
or zero-one loss versus the signs of entries in X, can be minimized.
Prior Work Previous results bounding the error of collaborative prediction using a lowrank matrix all assume the true target matrix Y is well-approximated by a low-rank matrix.
This corresponds to a large eigengap between the top few singular values of Y and the
remaining singular values. Azar et al [3] give asymptotic results on the convergence of
the predictions to the true preferences, assuming they have an eigengap. Drineas et al [4]
analyze the sample complexity needed to be able to predict a matrix with an eigengap, and
suggests strategies for actively querying entries in the target matrix. To our knowledge, this
is the first analysis of the generalization error of low-rank methods that do not make any
assumptions on the true target matrix.
Generalization error bounds (and related online learning bounds) were previously discussed
for collaborative prediction applications, but only when prediction was done for each user
separately, using a feature-based method, with the other user?s preferences as features [5,
6]. Although these address a collaborative prediction application, the learning setting is a
standard feature-based setting. These methods are also limited, in that learning must be
performed separately for each user.
Shaw-Taylor et al [7] discuss assumption-free post-hoc bounds on the residual errors of
low-rank approximation. These results apply to a different setting, where a subset of the
rows are fully observed, and bound a different quantity?the distance between rows and
the learned subspace, rather then the distance to predicted entries.
Organization In Section 2 we present a generalization error bound for zero-one loss,
based on a combinatorial result which we prove in Section 3. In Section 4 we generalize
the bound to arbitrary loss functions. Finally, in Section 5 we justify the combinatorial
approach taken, by considering an alternate approach (viewing rank-k matrices as combination of k rank-1 matrices) and showing why it does not work.
2
Generalization Error Bound for Zero-One Error
We begin by considering binary labels Yia ? ? and a zero-one sign agreement loss:
loss? (Xia ; Yia ) = 1Yia Xia ?0
(1)
Theorem 1. For any matrix Y ? {?1}n?m , n, m > 2, ? > 0 and integer k, with probability at least 1 ? ? over choosing a subset S of entries in Y uniformly among all subsets
of |S| entries, the discrepancy with respect to the zero-one sign agreement loss satisfies2 :
s
k(n + m) log 16em
? log ?
k
?X,rank X<k D? (X; Y ) < D?
(X;
Y
)
+
S
2|S|
To prove the theorem we employ standard arguments about the generalization error for
finite hypothesis classes with bounded cardinality.
First fix Y as well as X ? Rn?m . When an index pair (i, a) is chosen uniformly at random,
loss(Xia ; Yia ) is a Bernoulli random variable with probability D? (X; Y ) of being one.
If the entries of S are chosen independently and uniformly, |S|D?
S (X; Y ) is Binomially
distributed with mean |S|D? (X; Y ) and using Chernoff?s inequality:
?2|S|2
Pr D? (X; Y ) ? D?
(2)
S (X; Y ) + ? e
S
The distribution of S in Theorem 1 is slightly different, as S is chosen without repetitions.
The mean of D?
S (X; Y ) is the same, but it is more concentrated, and (2) still holds.
Now consider all rank-k matrices. Noting that loss(Xia ; Yia ) depends only on the sign
of Xia , it is enough to consider the equivalence classes of matrices with the same sign
patterns. Let f (n, m, k) be the number of such equivalence classes, i.e. the number of
possible sign configurations of n ? m matrices of rank at most k:
F (n, m, k) = {sign X ? {?, 0, +}n?m |X ? Rn?m , rank X ? k}
f (n, m, k) = ]F (n, m, k)
(
where sign X denotes the element-wise sign matrix (sign X)ia =
1
0
?1
If Xia > 0
If Xia = 0 .
If Xia < 1
For all matrices in an equivalence class, the random variable D?
S (X; Y ) is the same, and
taking a union bound of the events D? (X; Y ) ? D?
(X;
Y
)+
for
each of these f (n, m, k)
S
random variables we have:
s
!
log f (n, m, k) ? log ?
?
?
? ? (3)
Pr ?X,rank X?k D (X; Y ) ? DS (X; Y ) +
S
2|S|
q
?
by using (2) and setting = log f (n,m,k)?log
. The proof of Theorem 1 rests on bounding
2|S|
f (n, m, k), which we will do in the next section.
Note that since the equivalence classes we defined do not depend on the sample set, no
symmetrization argument is necessary.
2
All logarithms are base two
3
Sign Configurations of a Low-Rank Matrix
In this section, we bound the number f (n, m, k) of sign configurations of n ? m rankk matrices over the reals. Such a bound was previously considered in the context of
unbounded error communication complexity. Alon, Frankl and R?odl [8] showed that
(n+m)k+h+m
f (n, m, k) ? minh (8dnm/he)
, and used counting arguments to establish
that some (in fact, most) binary matrices can only be realized by high-rank matrices, and
therefore correspond to functions with high unbounded error communication complexity.
Here, we follow a general course outlined by Alon [9] to obtain a simpler, and slightly
tighter, bound based on the following result due to Warren:
Let P1 , . . . , Pr be real polynomials in q variables, and let C be the complement of the
variety defined by ?i Pi , i.e. the set of points in which all the m polynomials are non-zero:
C = {x ? Rq |?i Pi (x) 6= 0}
Theorem 2 (Warren [10]). If all r polynomials are of degree at most d, then the number
of connected components of C is at most:
q
q
X
r
4edr
c(C) ? 2(2d)q
2i
?
i
q
i=0
where the second inequality holds when r > q > 2.
The signs of the polynomials P1 , . . . , Pr are fixed inside each connected component of C.
And so, c(C) bounds the number of sign configurations of P1 , . . . , Pr that do not contain
zeros. To bound the overall number of sign configurations the polynomials are modified
slightly (see Appendix), yielding:
Corollary 3 ([9, Proposition 5.5]). The number of -/0/+ sign configurations of r polynomials, each of degree at most d, over q variables, is at most (8edr/q)q (for r > q > 2).
In order to apply these bounds to low-rank matrices, recall that any matrix X of rank at
most k can be written as a product X = U V 0 where U ? Rn?k and V ? Rk?m . Consider
the k(n+m) entries of U, V as variables, and the nm entries of X as polynomials of degree
two over these variables:
k
X
Ui? Va?
Xia =
?=1
Applying Corollary 3 we obtain:
k(n+m)
8e?2?nm
? (16em/k)k(n+m)
Lemma 4. f (n, m, k) ? k(n+m)
Substituting this bound in (3) establishes Theorem 1. The upper bound on f (n, m, k) is
tight up to a multiplicative factor in the exponent:
1
Lemma 5. For m > k 2 , f (n, m, k) ? m 2 (k?1)n
Proof. Fix any matrix V ? Rm?k with rows in general position, and consider the number
f (n, V, k) of sign configurations of matrices U V 0 , where U varies over all n ? k matrices.
Focusing only on +/? sign configurations (no zeros in U V 0 ), each row of sign U V 0 is a
homogeneous linear classification
of the
rows of V , i.e. of m vectors in general position
P
k?1
in Rk . There are exactly 2 i=0 mi
possible homogeneous linear classifications of m
vectors in general position in Rk , and so these many options for each row of sign U V 0 . We
can therefore bound:
!n
k?1
n(k?1)
X
1
m
m n
m
f (n, m, k) ? f (n, V, k) ? 2
?
?
= m 2 (k?1)n
i
k?1
k?1
i=0
4
Generalization Error Bounds for Other Loss Functions
In Section 2 we considered generalization error bounds for a zero-one loss function. More
commonly, though, other loss functions are used, and it is desirable to obtain generalization
error bounds for general loss functions.
When dealing with other loss functions, the magnitude of the entries in the matrix are
important, and not only their signs. It is therefore no longer enough to bound the number
of sign configurations. Instead, we will bound not only the number of ways low rank
matrices behave with regards to a threshold of zero, but the number of possible ways lowrank matrices can behave relative to any set of thresholds. That is, for any threshold matrix
T ? Rn?m , we will show that the number of possible sign configurations of (X ? T ),
where X is low-rank, is small. Intuitively, this captures the complexity of the class of
low-rank matrices not only around zero, but throughout all possible values.
We then use standard results from statistical machine learning to obtain generalization error
bounds from the bound on the number of relative sign configurations. The number of relative sign configurations serves as a bound on the pseudodimension?the maximum number
of entries for which there exists a set of thresholds such that all relative sign configurations
(limited to these entries) is possible. The pseudodimension can in turn be used to show the
existence of a small -net, which is used to obtain generalization error bounds.
Recall the definition of the pseudodimension of a class of real-valued functions:
Definition 1. A class F of real-valued functions pseudo-shatters the points x1 , . . . , xn
with thresholds t1 , . . . , tn if for every binary labeling of the points (s1 , . . . , sn ) ? {+, ?}n
there exists f ? F s.t. f (xi ) ? ti iff si = ?. The pseudodimension of a class F is the
supremum over n for which there exist n points and thresholds that can be shattered.
In order to apply known results linking the pseudodimension to covering numbers, we
consider matrices X ? Rn?m as real-valued functions X : [n] ? [m] ? R over index
pairs to entries in the matrix. The class Xk of rank-k matrices can now be seen as a class
of real-valued functions over the domain [n] ? [m]. We bound the pseudodimension of
this class by bounding, for any threshold matrix T ? Rn?m the number of relative sign
matrices:
FT (n, m, k) = {sign (X ? T ) ? {?, 0, +}n?m |X ? Rn?m , rank X ? k}
fT (n, m, k) = ]FT (n, m, k)
Lemma 6. For any T ? Rn?m , we have fT (n, m, k) ?
16em k(n+m)
.
k
Proof. We take a similar approach to that of Lemma 4, writing rank-k matrices as a product
X = U V 0 where U ? Rn?k and V ? Rk?m . Consider the k(n + m) entries of U, V as
variables, and the nm entries of X ? T as polynomials of degree two over these variables:
(X ? T )ia =
k
X
Ui? Va? ? Tia
?=1
Applying Corollary 10 yields the desired bound.
Corollary 7. The pseudodimension of the class Xk of n ? m matrices over the reals of
rank at most k, is at most k(n + m) log 16em
k .
We can now invoke standard generalization error bounds in terms of the pseudodimension
(Theorem 11 in the Appendix) to obtain:
Theorem 8. For any monotone loss function with |loss| ? M , any matrix Y ? {?1}n?m ,
n, m > 2, ? > 0 and integer k, with probability at least 1 ? ? over choosing a subset S of
entries in Y uniformly among all subsets of |S| entries:
v
u
u k(n + m) log 16em log M |S| ? log ?
t
k
k(n+m)
?X,rank X<k D(X; Y ) < DS (X; Y ) + 6
|S|
5
Low-Rank Matrices as Combined Classifiers
Rank-k matrices are those matrices which are a sum of k rank-1 matrices. If we view
matrices as functions from pairs of indices to the reals, we can think of rank-k matrices as
?combined? classifiers, and attempt to bound their complexity as such, based on the low
complexity of the ?basis? functions, i.e. rank-1 matrices.
A similar approach is taken in related work on learning with low-norm (maximum margin)
matrix factorization [11, 12], where the hypothesis class can be viewed as a convex combination of rank-1 unit-norm matrices. Scale-sensitive (i.e. dependent on the margin, or the
slope of the loss function) generalization error bounds for this class are developed based on
the graceful behavior of scale-sensitive complexity measures (e.g. log covering numbers
and the Rademacher complexity) with respect to convex combinations. Taking a similar
view, it is possible to obtain scale-sensitive generalization error bounds for low-rank matrices. In this Section we question whether it is possible to obtain scale-insensitive bounds,
similar to Theorems 1 and 8, by viewing low-rank matrices as combined classifiers.
It cannot be expected that scale-insensitive complexity would be preserved when taking
convex combinations of an unbounded number of base functions. However, the VCdimension, a scale-insensitive measure of complexity, does scale gracefully when taking
linear combinations of a bounded number of functions from a low VC-dimension class of
indicator function. Using this, we can obtain generalization error bounds for linear combinations of signs of rank-one matrices, but not signs of linear combinations of rank-one
matrices. An alternate candidate scale-insensitive complexity measure is the pseudodimension of a class of real-valued functions. If we could bound the pseudodimension of the
class of sums of k functions from a bounded-pseudodimension base class of real valued
functions, we could avoid the sign-configuration counting and obtain generalization error
bounds for rank-k matrices. Unfortunately, the following counterexample shows that this
is not possible.
Theorem 9. There exists a family F closed under scalar multiplication whose pseudodimension is at most five, and such that {f1 + f2 |f1 , f2 ? F} does not have a finite pseudodimension.
Proof. We describe a class F of real-valued functions over the positive integers N. To do
so, consider a one-to-one mapping of finite sets of positive integers to the positive integers.
For each A ? N define two functions3 , fA (x) = 2xA + 1x?A and gA (x) = 2xA . Let F be
the set of all scalar multiplications of these functions.
For every A ? N , fA ? gA is the indicator function of A, implying that every finite subset
can be shattered, and the pseudodimension of {f1 + f2 : f1 , f2 ? F} is unbounded.
It remains to show that the pseudodimension of F is less than six. To do so, we note that
there are no positive integers A < B and x < y and positive reals ?, ? > 0 such that
?(2xB + 1) > ?2xA and ?2yB < ?(2yA + 1). It follows that for any A < B and any
?, ? > 0, on an initial segment (possibly empty) of N we have ?gB ? ?fB ? ?gA ? ?fA
while on the rest of N we have ?gA ? ?fA < ?gB ? ?fB . In particular, any pair of
3
We use A to refer both to a positive integer and the finite set it maps to.
functions (?fA , ?fB ) or (?fA , ?gB ) or (?gA , ?gB ) in F that are not associated with the
same subset (i.e. A 6= B), cross each other at most once. This holds also when ? or ? are
negative, as the functions never change signs.
For any six naturals x1 < x2 < ? ? ? < x6 and six thresholds, consider the three labellings
(+, ?, +, ?, +, ?), (?, +, ?, +, ?, +), (+, +, ?, ?, +, +). The three functions realizing
these labellings must cross each other at least twice, but by the above arguments, there are
no three functions in F such that every pair crosses each other at least twice.4
6
Discussion
Alon, Frankl and R?odl [8] use a result of Milnor similar to Warren?s Theorem 2. Milnor?s
and Warren?s theorems were previously used for bounding the VC-dimension of certain
geometric classes [13], and of general concept classes parametrized by real numbers, in
terms of the complexity of the boolean formulas over polynomials used to represent them
[14]. This last general result can be used to bound the VC-dimension of signs of n?m rankk matrices by 2k(n + m) log(48enm), yielding a bound similar to Theorem 1 with an extra
log |S| term. In this paper, we take a simpler path, applying Warren?s theorem directly, and
thus avoiding the log |S| term and reducing the other logarithmic term. Applying Warren?s
theorem directly also enables us to bound the pseudodimension and obtain the bound of
Theorem 8 for general loss functions.
Another notable application of Milnor?s result, which likely inspired these later uses, is for
bounding the number of configurations of n points in Rd with different possible linear classifications [15, 16]. Viewing signs of rank-k n ? m matrices as n linear classification of m
points in Rk , this bound can be used to bound f (n, m, k) < 2km log 2n+k(k+1)n log n without using Warren?s Theorem directly [8, 12]. The bound of Lemma 4 avoids the quadratic
dependence on k in the exponent.
Acknowledgments We would like to thank Peter Bartlett for pointing out [13, 14]. N.S. and
T.J. would like to thank Erik Demaine for introducing them to oriented matroids.
A
Proof of Corollary 3
Consider a set R ? Rq containing one variable configuration for each possible sign pattern. Set
.
= 12 min1?i?q,x?RPi (x)6=0 |Pi (x)| > 0. Now consider the 2q polynomials Pi+ (x) = Pi (x) +
and Pi? (x) = Pi (x) ? and C 0 = x ? Rq |?i Pi+ (x) 6= 0, Pi? (x) 6= 0 . Different points in R
(representing all sign configurations) lie in different connected components of C 0 . Invoking Theorem
2 on C 0 establishes Corollary 3.
The count in Corollary 3 differentiates between positive, negative and zero signs. However, we are
only concerned with the positivity of Yia Xia (in the proof of Theorem 1) or of Xia ? Tia (in the
proof of Theorem 8), and do not need to differentiate
between zero and negative values. Invoking
Theorem 2 on C + = x ? Rq |?i Pi+ (x) 6= 0 , yields:
Corollary 10. The number of -/+ sign configurations (where zero is considered negative) of r polynomials, each of degree at most d, over q variables, is at most (4edr/q)q (for r > q > 2).
P
Applying Corollary 10 on the nm degree-two polynomials Yia k?=1 Ui? Va? establishes that for
any Y , the number of configurations of sign agreements of rank-k matrices with Y is bounded by
(8em/k)k(n+m) and yields a constant of 8 instead of 16 inside the logarithm in Theorem 1. Applying
Corollary 10 instead of Corollary 3 allows us to similarly tighten in the bounds in Corollary 7 and in
Theorem 8.
4
A more careful analysis shows that F has pseudodimension three.
B
Generalization Error Bound in terms of the Pseudodimension
Theorem 11. Let F be a class of real-valued functions f : X ? R with pseudodimension d, and
loss : R ? Y ? R be a bounded monotone loss function (i.e. for all y, loss(x, y) is monotone in x), with loss < M . For any joint distribution over (X, Y ), consider an i.i.d. sample
S = (X1 , Y1 ), . . . , (Xn , Yn ). Then for any > 0:
!
d
n
2 n
1X
32eM
Pr ?f ?F EX,Y [loss(f (X), Y )] >
loss(f (Xi ), Yi ) + < 4e(d + 1)
e? 32
S
n i=1
The bound is a composition of a generalization error bound in terms of the L1 covering number [17,
Theorem 17.1], a bound on the L1 covering number in terms of the pseudodimension [18] and the
observation that composition with a monotone function does not increase the pseudodimension [17,
Theorem 12.3].
References
[1] T. Hofmann. Latent semantic models for collaborative filtering. ACM Trans. Inf. Syst.,
22(1):89?115, 2004.
[2] Nathan Srebro and Tommi Jaakkola. Weighted low rank approximation. In 20th International
Conference on Machine Learning, 2003.
[3] Yossi Azar, Amos Fiat, Anna R. Karlin, Frank McSherry, and Jared Saia. Spectral analysis of
data. In ACM Symposium on Theory of Computing, pages 619?626, 2001.
[4] Petros Drineas, Iordanis Kerenidis, and Prabhakar Raghavan. Competitive recommendation
systems. In ACM Symposium on Theory of Computing, 2002.
[5] K. Crammer and Y. Singer. Pranking with ranking. In Advances in Neural Information Processing Systems, volume 14, 2002.
[6] Sanjoy Dasgupta, Wee Sun Lee, and Philip M. Long. A theoretical analysis of query selection
for collaborative filtering. Machine Learning, 51(3):283?298, 2003.
[7] John Shawe-Taylor, Nello Cristianini, and Jaz Kandola. On the concentration of spectral properties. In Advances in Neural Information Processing Systems, volume 14, 2002.
[8] N. Alon, P Frankl, and V. R?odel. Geometrical realization of set systems and probabilistic
communication complexity. In Foundations of Computer Science (FOCS), 1985.
[9] Noga Alon. Tools from higher algebra. In M. Gr?otschel R.L. Graham and L. Lov?asz, editors,
Handbook of Combinatorics, chapter 32, pages 1749?1783. North Holland, 1995.
[10] H. E. Warren. Lower bounds for approximation by nonlinear manifolds. Transactions of the
American Mathematical Society, 133:167?178, 1968.
[11] Nathan Srebro, Jason Rennie, and Tommi Jaakkola. Maximum margin matrix factorization. In
Advances in Neural Information Processing Systems, volume 17, 2005.
[12] Nathan Srebro. Learning with Matrix Factorization. PhD thesis, Massachusetts Institute of
Technology, 2004.
[13] Shai Ben-David and Michael Lindenbaum. Localization vs. identification of semi-algebraic
sets. Machine Learning, 32(3):207?224, 1998.
[14] Paul Goldberg and Mark Jerrum. Bounding the vapnik-chervonenkis dimension of concept
classes parameterized by real numbers. Machine Learning, 18(2-3):131?148, 1995.
[15] Jacob Goodman and Richard Pollack. Upper bounds for configurations and polytopes in Rd .
Discrete and Computational Geometry, 1:219?227, 1986.
[16] Noga Alon. The number of polytopes, configurations and real matroids. Mathematika, 33:62?
71, 1986.
[17] Martin Anthony and Peter L. Bartlett. Neural Networks Learning: Theoretical Foundations.
Cambridge University Press, 1999.
[18] David Haussler. Sphere packing numbers for subsets of the boolean n-cube with bounded
Vapnick-Chernovenkis dimension. J. Comb. Thoery, Ser. A, 69(2):217?232, 1995.
| 2700 |@word polynomial:12 norm:2 km:1 jacob:1 invoking:2 initial:1 configuration:21 chervonenkis:1 jaz:1 si:1 rpi:1 must:2 written:1 john:1 hofmann:1 enables:1 v:1 implying:1 intelligence:1 item:6 xk:2 realizing:1 matrix1:1 toronto:3 preference:8 simpler:2 five:1 unbounded:5 mathematical:2 symposium:2 focs:1 prove:3 fitting:1 inside:2 comb:1 lov:1 expected:1 behavior:1 p1:3 inspired:1 considering:2 cardinality:1 begin:1 bounded:6 israel:1 what:2 minimizes:1 developed:1 unobserved:1 pseudo:1 every:4 ti:1 exactly:1 rm:1 classifier:3 ser:1 unit:1 yn:1 t1:1 positive:7 influencing:1 dnm:1 path:1 might:1 twice:2 equivalence:4 suggests:1 ramat:1 limited:2 factorization:3 acknowledgment:1 union:1 refers:1 suggest:1 lindenbaum:1 cannot:1 ga:5 selection:1 context:1 applying:6 writing:1 map:1 independently:1 convex:3 haussler:1 his:1 target:5 user:10 homogeneous:2 us:1 goldberg:1 hypothesis:3 agreement:3 element:1 approximated:1 featurebased:1 observed:7 ft:4 min1:1 capture:1 connected:3 sun:1 rq:4 complexity:13 ui:3 cristianini:1 depend:1 tight:1 segment:1 algebra:1 localization:1 f2:4 basis:1 drineas:2 packing:1 joint:1 chapter:1 thoery:1 describe:1 artificial:1 query:1 labeling:1 choosing:2 pref:1 whose:1 valued:9 rennie:1 ability:1 jerrum:1 think:1 online:1 hoc:3 differentiate:1 net:1 karlin:1 product:2 realization:1 iff:1 convergence:1 empty:1 prabhakar:1 rademacher:1 ben:1 derive:1 alon:7 ac:1 vcdimension:1 school:1 lowrank:2 strong:1 c:1 predicted:3 indicate:1 tommi:4 differ:1 vc:3 raghavan:1 viewing:3 premise:1 fix:2 generalization:25 f1:4 proposition:1 tighter:1 hold:4 around:1 considered:3 mapping:1 predict:3 pointing:1 substituting:1 sought:1 label:2 combinatorial:2 symmetrization:1 sensitive:3 repetition:1 establishes:3 tool:1 weighted:1 amos:1 mit:1 modified:1 rather:1 pn:1 avoid:1 jaakkola:3 corollary:12 rank:41 likelihood:1 bernoulli:1 contrast:1 dependent:1 shattered:2 overall:2 among:2 classification:4 exponent:2 fairly:1 cube:1 once:1 never:1 chernoff:1 future:1 discrepancy:4 minimized:1 richard:1 few:2 employ:1 oriented:1 wee:1 kandola:1 geometry:1 attempt:1 organization:1 nogaa:1 interest:1 yielding:2 behind:1 mcsherry:1 xb:1 necessary:1 taylor:2 logarithm:2 desired:1 theoretical:2 pollack:1 boolean:2 introducing:1 entry:27 subset:10 predictor:1 gr:1 varies:1 combined:3 international:1 csail:1 probabilistic:4 invoke:1 pranking:1 lee:1 michael:1 thesis:1 squared:1 central:1 nm:5 containing:1 possibly:1 positivity:1 yia:9 american:1 actively:1 syst:1 tia:2 north:1 notable:1 combinatorics:1 ranking:1 depends:1 performed:1 multiplicative:1 view:2 closed:1 later:1 observing:1 analyze:1 jason:1 competitive:1 option:1 parallel:1 odel:1 shai:1 slope:1 collaborative:11 il:1 correspond:1 yield:3 generalize:1 identification:1 straight:1 vapnick:1 definition:2 associated:2 proof:7 mi:1 petros:1 massachusetts:2 recall:2 knowledge:1 fiat:1 focusing:1 higher:1 follow:1 x6:1 yb:1 done:1 though:1 erences:1 xa:3 d:5 nonlinear:1 aviv:2 usa:1 pseudodimension:23 contain:1 true:8 concept:2 laboratory:1 semantic:1 covering:4 criterion:1 tn:1 l1:2 geometrical:1 wise:1 insensitive:4 volume:3 discussed:1 he:1 linking:1 refer:1 composition:2 cambridge:2 counterexample:1 rd:2 outlined:1 pm:1 similarly:1 shawe:1 longer:1 base:3 showed:1 inf:1 certain:1 inequality:2 binary:3 yi:1 seen:2 semi:1 desirable:1 cross:3 long:1 sphere:1 justifying:1 post:3 y:1 va:3 prediction:16 represent:1 preserved:1 separately:2 singular:2 source:2 goodman:1 noga:3 rest:2 extra:1 asz:1 integer:7 noting:1 counting:2 enough:2 concerned:1 variety:1 whether:1 six:3 bartlett:2 gb:4 eigengap:3 peter:2 algebraic:1 rankk:2 concentrated:1 exist:1 frankl:3 sign:39 correctly:1 odl:2 discrete:1 dasgupta:1 threshold:8 drawn:1 shatters:1 monotone:4 sum:3 parameterized:1 throughout:1 family:1 appendix:2 graham:1 bound:61 correspondence:1 quadratic:1 x2:1 nathan:4 argument:4 graceful:1 martin:1 department:1 according:1 alternate:2 combination:7 slightly:3 em:7 labellings:2 making:1 s1:1 intuitively:1 pr:6 taken:2 previously:3 remains:1 discus:1 turn:1 count:1 differentiates:1 needed:1 yossi:1 jared:1 singer:1 serf:1 studying:2 apply:3 observe:1 spectral:2 shaw:1 existence:1 top:2 remaining:1 denotes:1 establish:1 society:1 objective:4 question:1 quantity:1 realized:1 enm:1 strategy:1 fa:6 dependence:2 concentration:1 traditional:1 subspace:1 distance:2 thank:2 otschel:1 parametrized:1 philip:1 gracefully:1 manifold:1 nello:1 assuming:1 erik:1 index:3 providing:1 unfortunately:1 relate:2 frank:1 negative:4 binomially:1 unknown:1 upper:2 observation:1 finite:6 minh:1 behave:2 communication:3 y1:1 rn:9 arbitrary:2 canada:1 rating:2 david:2 complement:1 pair:5 learned:1 polytopes:2 trans:1 address:1 able:4 pattern:2 tau:1 ia:3 overlap:1 event:1 natural:2 predicting:1 indicator:2 residual:1 representing:1 technology:2 rated:1 sn:1 prior:1 geometric:1 nati:1 dislike:1 multiplication:2 asymptotic:1 relative:5 loss:26 fully:1 filtering:4 demaine:1 srebro:4 versus:1 querying:1 foundation:2 degree:6 editor:1 pi:10 row:6 course:1 last:1 free:1 warren:8 institute:2 taking:4 matroids:2 distributed:1 regard:1 xia:13 dimension:5 xn:2 avoids:1 fb:3 forward:1 made:1 commonly:1 saia:1 far:1 tighten:1 transaction:1 supremum:1 dealing:1 handbook:1 xi:2 latent:1 why:1 nature:1 tel:1 anthony:1 domain:1 anna:1 accustomed:1 bounding:6 azar:2 paul:1 x1:3 position:3 candidate:1 lie:1 theorem:26 rk:5 formula:1 showing:1 exists:3 vapnik:1 phd:1 magnitude:1 margin:3 logarithmic:1 likely:1 partially:1 scalar:2 recommendation:1 holland:1 applies:1 corresponds:1 acm:3 ma:1 viewed:1 careful:1 change:1 determined:1 uniformly:5 edr:3 justify:1 useritem:1 reducing:1 lemma:5 sanjoy:1 ya:1 milnor:3 mark:1 crammer:1 avoiding:1 ex:1 |
1,874 | 2,701 | Algebraic Set Kernels with Application to
Inference Over Local Image Representations
Amnon Shashua and Tamir Hazan ?
Abstract
This paper presents a general family of algebraic positive definite similarity functions over spaces of matrices with varying column rank. The
columns can represent local regions in an image (whereby images have
varying number of local parts), images of an image sequence, motion trajectories in a multibody motion, and so forth. The family of set kernels
we derive is based on a group invariant tensor product lifting with parameters that can be naturally tuned to provide a cook-book of sorts covering
the possible ?wish lists? from similarity measures over sets of varying
cardinality. We highlight the strengths of our approach by demonstrating the set kernels for visual recognition of pedestrians using local parts
representations.
1
Introduction
In the area of learning from observations there are two main paths that are often mutually
exclusive: (i) the design of learning algorithms, and (ii) the design of data representations.
The algorithm designers take pride in the fact that their algorithm can generalize well given
straightforward data representations (most notable example is SVM [11]), whereas those
who work on data representations demonstrate often remarkable results with sophisticated
data representations using only straightforward learning algorithms (e.g. [5, 10, 6]). This
dichotomy is probably most emphasized in the area of computer vision, where image understanding from observations involve data instances of images or image sequences containing
huge amounts of data. A straightforward representation treating all the measurements as
a single vector, such as the raw pixel data, or a transformed raw-pixel data, places unreasonable demands on the learning algorithm. The ?holistic? representations suffer also
from sensitivity to occlusions, invariance to local and global transformations, non-rigidity
of local parts of the object, and so forth.
Practitioners in the area of data representations have long noticed that a collection of local
representations (part-based representations) can be most effective to ameliorate changes of
appearance [5, 10, 6]. The local data representations vary in their sophistication, but share
the same principle where an image corresponds to a collection of points each in a relatively
small dimensional space ? instead of a single point in high-dimensional space induced
by holistic representations. In general, the number of points (local parts) per image may
vary and the dimension of each point may vary as well. The local representations tend
?
School of Engineering and Computer Science, Hebrew University of Jerusalem, Jerusalem
91904, Israel
to be robust against occlusions, local and global transformations and preserve the original
resolution of the image (the higher the resolution the more parts are generated per image).
The key for unifying local and holistic representations for inference engines is to design
positive definite similarity functions (a.k.a. kernels) over sets (of vectors) of varying cardinalities. A Support Vector Machine (SVM) [11] can then handle sets of vectors as a single
instance via application of those ?set kernels?. A set kernel would be useful also to other
types of inference engines such as kernel versions of PCA, LDA, CCA, ridge regression and
any algorithm which can be mapped onto inner-products between pairs of data instances
(see [8] for details on kernel methods).
Formally, we consider an instance being represented by a collection of vectors, which for
the sake of convenience, form the columns of a matrix. We would like to find an algebraic
family of similarity functions sim(A, B) over matrices A, B which satisfy the following
requirements: (i) sim(A, B) is an inner product, i.e., sim(A, B) = ?(A)> ?(B) for some
mapping ?() from matrices to vectors, (ii) sim(A, B) is built over local kernel functions
k(ai , bj ) over columns ai and bj of A, B respectively, (iii) The column cardinality (rank
of column space) of A and B need not be the same (number of local parts may differ from
image to image), and (iv) the parameters of sim(A, B) should induce the properties of invariance to order (alignement) of parts, part occlusions, and degree of interactions between
local parts. In a nutshell, our work provides a cook-book of sorts which fundamentally
covers the possible algebraic kernels over collections of local representations built on top
of local kernels by combining (linearly and non-linearly) local kernels to form a family of
global kernels over local representations.
The design of a kernel over sets of vectors has been recently attracting much attention in the
computer vision and machine learning literature. A possible approach is to fit a distribution
to the set of vectors and define the kernel as a distribution matching measure [9, 12, 4].
This has the advantage that the number of local parts can vary but at the expense of fitting
a distribution to the variation over parts. The variation could be quite complex at times,
unlikely to fit into a known family of distributions in many situations of interest, and in
practice the sample size (number of columns of A) is not sufficiently large to reliably fit
a distribution. The alternative, which is the approach taken in this paper, is to create a
kernel over sets of vectors in a direct manner. When the column cardinality is equal it is
possible to model the similarity measure as a function over the principal angles between the
two column spaces ([14] and references therein) while for varying column cardinality only
heuristic similarity measures (which are not positive definite) have so far been introduced
[13].
It is important to note that although we chose SVM over local representations as the application to demonstrate the use of set kernels, the need for adequately working with instances
made out of sets of various cardinalities spans many other application domains. For example, an image sequence may be represented by a set (ordered or unordered) of vectors,
where each vector stands for an image, the pixels in an image can be represented as a tuple
consisting of position, intensity and other attributes, motion trajectories of multiply moving bodies can be represented as a collection of vectors, and so on. Therefore, the problem
addressed in this paper is fundamental both theoretically and from a practical perspective
as well.
2
The General Family of Inner-Products over Matrices
We wish to derive the general family of positive definite similarity measures sim(A, B)
over matrices A, B which have the same number of rows but possibly different column
rank (in particular, different number of columns). Let A be of dimensions n ? k and
B of dimension n ? q where n is fixed and k, q can vary at will over the application of
sim(?, ?) on pairs of matrices. Let m = max{n, k, q} be the upper bound over all values
of k, q encountered by the data. Let ai , bj be the column vectors of matrices A, B and
let k(ai , bj ) be the local kernel function. For example, in the context where the column
vectors represent local parts of an image, then the matching function k(?, ?) between pairs
of local parts provides the building blocks of the overall similarity function. The local
kernel is some positive definite function k(x, y) = ?(x)> ?(y) which is the inner-product
between the ?feature?-mapped vectors x, y for some feature map ?(?). For example, if ?(?)
is the polynomial map of degree up to d, then k(x, y) = (1 + x> y)d .
The local kernels can be combined in a linear or non-linear manner. When the combination
is linear the similarity becomes the analogue of the inner-product between vectors extended
to matrices. We will refer to the linear family as sim(A, B) =< A, B > and that will be
the focus of this section. In the next section we will derive the general (algebraic) nonlinear family which is based on ?lifting? the input matrices A, B onto higher dimensional
spaces and feeding the result onto the < ?, ? > machinery developed in this section, i.e.,
sim(A, B) =< ?(A), ?(B) >.
We will start by embedding A, B onto m ? m matrices by zero padding as follows. Let
ei denote the i?th standard basis vector (0, .., 0, 1, 0, .., 0) of Rm . The the embedding is
represented by linear combinations of tensor products:
A?
n X
k
X
aij ei ? ej ,
B?
q
n X
X
blt el ? et .
l=1 t=1
i=1 j=1
Note that A, B are the upper-left blocks of the zero-padded
matrices. Let S be a positive
Pp
semi definite m2 ? m2 matrix represented by S = r=1 Gr ? Fr where Gr , Fr are m ? m
? r be the n ? n
matrices1 . Let F?r be the q ? k upper-left sub-matrix of Fr> , and let G
upper-left sub-matrix of Gr . We will be using the following three identities:
Gx1 ? F x2 = (G ? F )(x1 ? x2 ),
(G ? F )(G0 ? F 0 ) = GG0 ? F F 0 ,
>
< x1 ? x2 , y1 ? y2 >= (x>
1 y1 )(x2 y2 ).
The inner-product < A, B > over all p.s.d. matrices S has the form:
X
X
X
< A, B > = <
aij ei ? ej , (
Gr ? Fr )
blt el ? et >
r
i,j
=
XX
r
XX
=
XX
r
=
>
aij blt (e>
i Gr el )(ej Fr et )
i,j,l,t
aij blt (Gr )il (Fr )jt
i,j,l,t
XX
r
aij blt < ei ? ej , Gr el ? Fr et >
i,j,l,t
=
r
l,t
? r B)jt (Fr )jt
(A> G
lt
!
= trace
X
? r B)F?r
(A G
>
r
We have represented the inner product < A, B > using the choice of m ? m matrices
Gr , Fr instead of the choice of a single m2 ? m2 p.s.d. matrix S. The matrices Gr , Fr
1
Any S can be represented as a sum over tensor products: given column-wise ordering, the matrix
G ? F is composed of n ? n blocks of the form fij G. Therefore, take Gr to be the n ? n blocks of
S and Fr to be the elemental matrices which have ?1? in coordinate r = (i, j) and zero everywhere
else.
Pp
must be selected such that r=1 Gr ? Fr is positive semi definite. The problem of deciding on the the necessary conditions on Fr and Gr such that the sum over tensor products is
p.s.d is difficult. Even deciding whether a given S has a separable decomposition is known
to be NP-hard [3]. The sufficient
conditions are easy ? choosing Gr , Fr to be positive
Pp
semi definite would make r=1 Gr ? Fr positive semi definite as well. In this context
(of separable S) we need one more constraint in order to work with non-linear local ker?r = M
? r> M
? r must ?distribute with the kernel?,
nels k(x, y) = ?(x)> ?(y): the matrices G
namely there exist Mr such that
? >M
? r ?(y) = ?(x)> G
? r ?(y).
k(Mr x, Mr y) = ?(Mr x)> ?(Mr y) = ?(x)> M
r
To summarize the results so far, the most general, but seperable, analogue of the innerproduct over vectors to the inner-product of matrices of varying column cardinality has the
form:
X
< A, B >=
trace(Hr F?r )
(1)
r
Where the entries of Hr consists of k(Mr ai , Mr bj ) over the columns of A, B after possibly
? r ), and F?r are the q ? k upperundergoing global coordinate changes by Mr (the role of G
>
left sub-matrix of positive definite m ? m matrices Fr .
? r is to perform global coordinate changes of Rn before applicaThe role of the matrices G
tion of the kernel k() on the columns of A, B. These global transformations include projections (say onto prototypical ?parts?) that may be given or ?learned? from a training set.
The matrices F?r determine the range of interaction between columns of A and columns of
? r = I then < A, B >= trace(A> B F? ) where F? is the upper-left
B. For example, when G
P
submatrix with the appropriate dimension of some fixed m ? m p.s.d matrix F = r Fr .
Note that entries of A> B are k(ai , bj ). In other words, when
P Gr = I, < A, B > boils
down to a simple linear super-position of the local kernels, ij k(ai , bj )fij where the entries fij are part of the upper-left block of a fixed positive definite matrix F where the
block dimensions are commensurate with the number of columns of A and those of B. The
various choices of F determine the type of invariances one could obtain from the similarity measure. For example, when F = I the similarity is simply the sum (average) of
the local kernels k(ai , bi ) thereby assuming we have a strict alignment between the local
parts represented by A and the local parts represented by B. On the other end of the invariance spectrum, when F = 11> (all entries are ?1?) the similarity measure averages
over all interactions of local parts k(ai , bj ) thereby achieving an invariance to the order of
the parts. A decaying weighted interaction such as fij = ? ?|i?j| would provide a middle
ground between the assumption of strict alignment and the assumption of complete lack of
alignment. In the section below we will derive the non-linear version of sim(A, B) based
on the basic machinery of < A, B > of eqn. (1) and lifting operations on A, B.
3
Lifting Matrices onto Higher Dimensions
The family of sim(A, B) =< A, B > forms a weighted linear superposition of the
local kernel k(ai , bj ). Non-linear combinations of local kernels emerge using mappings ?(A) from the input matrices onto other higher-dimensional matrices, thus forming
sim(A, B) =< ?(A), ?(B) >. Additional invariance properties and parameters controlling the perfromance of sim(A, B) emerge with the introduction of non-linear combinations of local kernels, and those will be discussed later on in this section.
Consider the general d-fold lifting ?(A) = A?d which can be viewed as a nd ? k d matrix.
Let Fr be a p.s.d. matrix of dimension md ? md and F?r be the upper-left q d ? k d block of
? r )?d be a p.s.d matrix of dimension nd ? nd where G
? r is p.s.d. n ? n
Fr . Let Gr = (G
matrix. Using the identity (A?d )> B ?d = (A> B)?d we obtain the inner-product in the
lifted space:
< A?d , B ?d >=
X
? r B)?d F?r .
trace (A> G
r
?l
By taking linear combinations of < A , B ?l >, l = 1, ..., d, we get the general nonhomogenous d-fold inner-product simd (A, B). A this point the formulation is general but
somewhat unwieldy computational-wise. The key for computational simplification lay in
the fact that choices of Fr determine not only local interactions (as in the linear case) but
also group invariances. The group invariances are a result of applying symmetric operators
on the tensor product space ? we will consider two of those operators here, known as the
the d-fold alternating tensor A?d = A ? .... ? A and the d-fold symmetric tensor A?d =
A ? ... ? A. These lifting operations introduce the determinant and permanent operations on
? r B, as described below.
submatrices of A> G
The alternating tensor is a multilinear map of Rn , (A ? .... ? A)(x1 ? ... ? xd ) = Ax1 ?
... ? Axd , where
1 X
x1 ? ... ? xd =
sign(?)x?(1) ? .... ? x?(d) ,
d!
??Sd
where Sd is the symmetric group over d letters and ? ? Sd are the permutations of the
group. If x1 , ..., xn form a basis of Rn , then the nd elements xi1 ? ... ? xid , where 1 ?
i1 < ... < id ? n form a basis of the alternating d ? f old tensor product of Rn , denoted
as ?d Rn . If A ? Rn?k is a linear map on Rn sending points to Rk , then A?d is a
linear map on ?d Rn sending x1 ? ... ? xd to Ax1 ? ... ? Axd , i.e., sending points in
?d Rn to points in ?d Rk . The matrix representation of A?d is called the ?d?th compound
matrix? Cd (A) whose (i1 , ..., id |j1 , ..., jd ) entry has the value det(A[i1 , ..., id : j1 , ..., jd ])
where the determinant is of the d ? d block constructed by choosing
the rows
i1 , ..., id
and the columns j1 , ..., jd of A. In other words, Cd (A) has nd rows and kd columns
(instead of nd ? k d necessary for A?d ) whose entries are equal to the d ? d minors of A.
When k = d, Ck (A) is a vector known as the Grasmanian of A, and when n = k = d
then Cd (A) = det(A). Finally, the identity (A?d )> B ?d = (A> B)?d specializes to
(A?d )> B ?d = (A> B)?d which translates to the identity Cd (A)> Cd (B) = Cd (A> B)
known as the Binet-Cauchy theorem [1]. Taken together, the ?d-fold alternating kernel?
?d (A, B) is defined by:
X
? r B)F?r , (2)
?d (A, B) =< A?d , B ?d >=< Cd (A), Cd (B) >=
trace Cd (A> G
r
m
where F?r is the
?
upper-left submatrix of the p.s.d m
d ? d matrix Fr . Note
? r B)ij = k(Mr ai , Mr bj ) where G
?r =
that the local kernel plugs in as the entries of (A> G
>
Mr Mr .
q
d
k
d
Another symmetric operator on the tensor product space is via the d-fold symmetric tensor
space Symd Rn whose points are:
1 X
x?(1) ? .... ? x?(d) .
x1 ? ? ? xd =
d!
??Sd
The analogue of Cd (A) is the ?d?th power matrix? Rd (A) whose (i1 , ..., id |j1 , ..., jd ) entry
has the value perm(A[i1 , ..., id : j1 , ..., jd ]) and which stands for the map A?d
(A ? ? ? A)(x1 ? ? ? xd ) = Ax1 ? ? ? Axd .
In other words, Rd (A) has n+d?1
rows and k+d?1
columns whose entries are equal to
d
d
the d?d permanents of A. The analogue of the Binet-Cauchy theorem is Rd (A)> Rd (B) =
Rd (A> B). The ensuing kernel similarity function, referred to as the ?d-fold symmetric
kernel? is:
X
? r B)F?r (3)
Symd (A, B) =< A?d , B ?d >=< Rd (A), Rd (B) >=
trace Rd (A> G
r
k+d?1
where F?r is the q+d?1
?
upper-left submatrix of the positive definite m+d?1
?
d
d
d
n+d?1
matrix
F
.
Due
to
lack
of
space
we
will
stop
here
and
spend
the
remainder
of
this
r
d
section in describing in laymen terms what are the properties of these similarity measures,
how they can be constructed in practice and in a computationally efficient manner (despite
the combinatorial element in their definition).
3.1 Practical Considerations
To recap, the family of similarity functions sim(A, B) comprise of the linear version
< A, B > (eqn. 1) and non-linear versions ?l (A, B), Syml (A, B) (eqns. 2,3) which are
group projections of the general kernel < A?d , B ?d >. These different similarity functions are controlled by the choice of three items: Gr , Fr and the parameter d representing
the degree of the tensor product operator. Specifically, we will focus on the case Gr = I
? r is fairly inand on ?d (A, B) as a representative of the non-linear family. The role of G
teresting as it can be viewed as a projection operator from ?parts? to prototypical parts that
can be learned from a training set but we leave this to the full length article that will appear
later.
Practically, to compute ?d (A, B) one needs to run over all d ? d blocks of the k ? q matrix A> B (whose entries are k(ai , bj )) and for each block compute the determinant. The
similarity function is a weighted sum of all those determinants weighted by fij . By appropriate selection of F one can control both the complexity (avoid running over all possible
d ? d blocks) of the computation and the degree of interaction between the determinants.
These determinants have an interesting geometric interpretation if those are computed over
unitary matrices ? as described next.
Let A = QA RA and B = QB RB be the QR factorization of the matrices, i.e., QA has
orthonormal columns which span the column space of A, then it has been recently shown
?1
[14] that RA
can be computed from A using only operations over k(ai , aj ). Therefore,
?T >
?1
the product Q>
A QB , which is equal to RA A BRB , can be computed using only local
?1
kernel applications. In other words, for each A compute RA
(can be done using only
inner-products over columns of A), then when it comes to compute A> B compute in?T >
?1
stead RA
A BRB
which is equivalent to computing Q>
A QB . Thus effectively we have
replaced every A with QA (unitary matrix).
Now, ?d (QA , QB ) for unitary matrices is the sum over the product of the cosine principal
angles between d-dim subspaces spanned by columns of A and B. The value of each
determinant of the d ? d blocks of Q>
A QB is equal to the product of the cosine principal
angles between the respective d-dim subspaces determined by corresponding selection of
d columns from A and d columns from B. For example, the case k = q = d produces
>
?d (QA , QB ) = det(Q>
A QB ) which is the product of the eigenvalues of the matrix QA QB .
Those eigenvalues are the cosine of the principal angles between the column space of A
and the column space of B [2]. Therefore, det(Q>
A QB ) measures the ?angle? between the
two subspaces spanned by the respective columns of the input matrices ? in particular is
invariant to the order of the columns. For smaller values of d we obtain the sum over such
products between subspaces spanned by subsets of d columns between A and B.
The advantage of smaller values of d is two fold: first it enables to compute the similarity
when k 6= q and second breaks down the similarity between subspaces into smaller pieces.
The entries of the matrix F determine which subspaces are being considered and the interaction between subspaces in A and B. A diagonal F compares corresponding subspaces
(a)
(b)
Figure 1: (a) The configuration of the nine sub-regions is displayed over the gradient image. (b)
some of the positive examples ? note the large variation in appearance, pose and articulation.
between A and B whereas off-diagonal entries would enable comparisons between different choices of subspaces in A and in B. For example, we may want to consider choices
of d columns arranged in a ?sliding? fashion, i.e., column sets {1, .., d}, {2, ..., d + 1}, ...
and so forth, instead of the combinatorial number of all possible choices. This selection
is associated with a sparse diagonal F where the non-vanishing entries along the diagonal
have the value of ?1? and correspond to the sliding window selections.
To conclude, in the linear version < A, B > the role of F is to determine the range of
interaction between columns of A and columns of B, whereas with the non-linear version
it is the interaction between d-dim subspaces rather than individual columns. We could
select all possible interactions (exponential number) or any reduced interaction set such as
the sliding window rule (linear number of choices) as described above.
4
Experiments
We examined the performance of sim(A, B) on part-based representations for pedestrian
detection using SVM for the inference engine. The dataset we used (courtesy of Mobileye
Ltd.) covers a challenging variability of appearance, viewing position and body articulation
(see Fig. 1). We ran a suit of comparative experiments using sim(A, B) =< A, B >
with three versions of F = {I, 11> , decay} with local kernels covering linear, d?th degree
polynomial (d = 2, 6) and RBF kernel, and likewise with sim(A, B) = ?d (A, B) with d =
2, sparse diagonal F (covering a sliding window configuration) and with linear, polynomial
and RBF local kernels. We compared our results to the conventional down-sampled holistic
representation where the raw images were down-sampled to size 20 ? 20 and 32 ? 32.
Our tests also included simulation of occlusions (in the test images) in order to examine
the sensitivity of our sim(A, B) family to occlusions. For the local part representation,
the input image was divided into 9 fixed regions where for each image local orientation
statistics were were generated following [5, 7] with a total of 22 numbers per region (see
Fig 1a), thereby making a 22 ? 9 matrix representation to be fed into sim(A, B). The
size of the training set was 4000 split evenly between positive and negative examples and
a test set of 4000 examples was used to evaluate the performance of each trial. The table
below summarizes the accuracy results for the raw-pixel (holistic) representation over three
trials: (i) images down-sampled to 20 ? 20, (ii) images down-sampled to 32 ? 32, and
(iii) test images were partially occluded (32 ? 32 version). The accuracy figures are the
ratio between the sum of the true positives and true negatives and the total number of test
examples.
20 ? 20
32 ? 32
occlusion
raw linear
78%
78%
73.5%
poly d = 2
83%
84%
72%
poly d = 6
84%
85%
77%
RBF
86%
82%
76.5%
The table below displays sim(A, B) with linear and RBF local kernels.
local kernel
linear
RBF
< A, B >, F = I
< A, B >, F = 11>
< A, B >, fij = 2?|i?j|
?2 (A, B)
90.8%
91.2%
85%
85%
90.6%
90.4%
88%
90%
One can see that the local part representation provides a sharp increase in accuracy compared to the raw pixel holistic representation. The added power of invariance to order of
parts induced by < A, B >, F = 11> is not required since the parts are aligned and therefore the accuracy is the highest for the linear combination of local RBF < A, B >, F = I.
The same applies for the non-linear version ?d (A, B) ? the additional invariances that
come with a non-linear combination of local parts are apparently not required. The power
of non-linearity associated with the combination of local parts comes to bear when the
test images have occluded parts, i.e., at random one of the columns of the input matrix is
removed (or replaced with a random vector), as shown in the table below:
local kernel
linear
RBF
< A, B >, F = I
?2 (A, B)
62%
83%
87%
88%
One can notice that a linear combination of local parts suffers from reduced accuracy
whereas the non-linear combination maintains a stable accuracy (compare the right-most
columns of the two tables above). Although the experiments above are still preliminary
they show the power and potential of the sim(A, B) family of kernels defined over local
kernels. With the principles laid down in Section 3 one can construct a large number (we
touched only a few) of algebraic kernels which combine the local kernels in non-linear
ways thus creating invariances to order and increased performance against occlusion. Further research is required for sifting through the various possibilities with this new family
of kernels and extracting their properties, their invariances and behavior under changing
parameters (Fr , Gr , d).
References
[1] A.C. Aitken. Determinants and Matrices. Interscience Publishers Inc., 4th edition, 1946.
[2] G.H. Golub and C.F. Van Loan. Matrix computations. John Hopkins University Press, 1989.
[3] L. Gurvits. Classical deterministic complexity of edmonds? problem and quantum entanglement. In ACM Symp. on Theory of Computing, 2003.
[4] R. Kondor and T. Jebara. A kernel between sets of vectors. In International Conference on
Machine Learning, ICML, 2003.
[5] D.G. Lowe. Distinctive image features from scale-invariant keypoints. International Journal of
Computer Vision, 2004.
[6] C. Schmidt and R. Mohr. Local grey-value invariants for image retrieval. IEEE Transactions
on Pattern Analysis and Machine Intelligence, 19(5):530?535, 1997.
[7] A. Shashua, Y. Gdalyahu, G. Hayun and L. Mann. ?Pedestrian Detection for Driving Assistance
Systems?. IEEE Intelligent Vehicles Symposium (IV2004), June. 2004, Parma Italy.
[8] B. Sch?olkopf and A.J. Smola. Learning with Kernels. MIT Press, Cambridge, MA, 2002.
[9] G. Shakhnarovich, J.W. Fisher, and T. Darrell. Face recognition from long-term observations.
In Proceedings of the European Conference on Computer Vision, 2002.
[10] S. Ullman, M. Vidal-Naquet, and E. Sali. Visual features of intermediate complexity and their
use in classification. Nature Neuroscience, 5(7):1?6, 2002.
[11] V.N. Vapnik. The nature of statistical learning. Springer, 2nd edition, 1998.
[12] N. Vasconcelos, P. Ho, and P. Moreno. The kullback-leibler kernel as a framework for discriminant and localized representations for visual recognition. In Proceedings of the European
Conference on Computer Vision, pages 430?441, Prague, Czech Republic, May 2004.
[13] C. Wallraven, B. Caputo, and A. Graf. Recognition with local features: the kernel recipe. In
Proceedings of the International Conference on Computer Vision, 2003.
[14] L. Wolf and A. Shashua. Learning over sets using kernel principal angles. Journal of Machine
Learning Research (JMLR), 4(10):913?931, 2003.
| 2701 |@word trial:2 determinant:8 middle:1 version:9 polynomial:3 kondor:1 nd:7 grey:1 simulation:1 decomposition:1 thereby:3 configuration:2 tuned:1 must:2 john:1 j1:5 enables:1 moreno:1 treating:1 intelligence:1 selected:1 cook:2 item:1 vanishing:1 provides:3 along:1 constructed:2 direct:1 symposium:1 consists:1 fitting:1 combine:1 interscience:1 symp:1 introduce:1 manner:3 theoretically:1 aitken:1 ra:5 behavior:1 examine:1 window:3 cardinality:7 becomes:1 multibody:1 xx:4 linearity:1 israel:1 what:1 developed:1 transformation:3 every:1 xd:5 nutshell:1 rm:1 control:1 appear:1 positive:15 before:1 engineering:1 local:54 sd:4 despite:1 id:6 path:1 mohr:1 chose:1 therein:1 examined:1 challenging:1 factorization:1 range:2 bi:1 gdalyahu:1 practical:2 practice:2 block:12 definite:12 ker:1 area:3 ax1:3 submatrices:1 matching:2 projection:3 word:4 induce:1 get:1 onto:7 convenience:1 selection:4 operator:5 context:2 applying:1 equivalent:1 map:6 conventional:1 courtesy:1 deterministic:1 straightforward:3 jerusalem:2 attention:1 resolution:2 m2:4 rule:1 orthonormal:1 spanned:3 embedding:2 handle:1 variation:3 coordinate:3 controlling:1 element:2 recognition:4 lay:1 role:4 region:4 ordering:1 sifting:1 highest:1 removed:1 ran:1 entanglement:1 complexity:3 occluded:2 shakhnarovich:1 gg0:1 distinctive:1 basis:3 wallraven:1 represented:10 various:3 effective:1 dichotomy:1 choosing:2 quite:1 heuristic:1 seperable:1 whose:6 spend:1 say:1 stead:1 statistic:1 sequence:3 advantage:2 eigenvalue:2 interaction:11 product:24 fr:23 remainder:1 aligned:1 combining:1 holistic:6 forth:3 olkopf:1 qr:1 recipe:1 elemental:1 requirement:1 darrell:1 produce:1 comparative:1 leave:1 object:1 derive:4 pose:1 ij:2 minor:1 school:1 sim:21 matrices1:1 come:3 differ:1 fij:6 attribute:1 enable:1 viewing:1 mann:1 xid:1 feeding:1 preliminary:1 multilinear:1 practically:1 sufficiently:1 recap:1 ground:1 considered:1 deciding:2 mapping:2 bj:11 driving:1 vary:5 combinatorial:2 superposition:1 create:1 weighted:4 mit:1 super:1 ck:1 rather:1 avoid:1 ej:4 lifted:1 varying:6 focus:2 june:1 rank:3 dim:3 inference:4 el:4 unlikely:1 transformed:1 i1:6 pixel:5 overall:1 classification:1 orientation:1 denoted:1 fairly:1 equal:5 simd:1 comprise:1 construct:1 gurvits:1 vasconcelos:1 icml:1 np:1 fundamentally:1 intelligent:1 few:1 composed:1 preserve:1 individual:1 replaced:2 occlusion:7 consisting:1 suit:1 detection:2 huge:1 interest:1 possibility:1 multiply:1 alignment:3 golub:1 tuple:1 necessary:2 respective:2 machinery:2 iv:1 old:1 instance:5 column:42 increased:1 cover:2 republic:1 entry:13 subset:1 gr:19 combined:1 fundamental:1 sensitivity:2 international:3 xi1:1 off:1 together:1 hopkins:1 containing:1 possibly:2 book:2 creating:1 ullman:1 distribute:1 potential:1 unordered:1 pedestrian:3 inc:1 satisfy:1 notable:1 permanent:2 piece:1 vehicle:1 tion:1 later:2 break:1 lowe:1 hazan:1 apparently:1 shashua:3 start:1 sort:2 decaying:1 maintains:1 il:1 accuracy:6 who:1 likewise:1 correspond:1 generalize:1 raw:6 trajectory:2 suffers:1 definition:1 against:2 pp:3 naturally:1 associated:2 boil:1 stop:1 sampled:4 dataset:1 sophisticated:1 higher:4 formulation:1 done:1 arranged:1 smola:1 working:1 eqn:2 ei:4 nonlinear:1 lack:2 aj:1 lda:1 building:1 y2:2 binet:2 true:2 adequately:1 alternating:4 symmetric:6 leibler:1 assistance:1 eqns:1 covering:3 whereby:1 cosine:3 ridge:1 demonstrate:2 complete:1 motion:3 image:29 wise:2 consideration:1 recently:2 blt:5 discussed:1 interpretation:1 measurement:1 refer:1 cambridge:1 ai:13 rd:8 axd:3 moving:1 stable:1 similarity:19 attracting:1 perspective:1 italy:1 compound:1 additional:2 somewhat:1 mr:12 layman:1 determine:5 ii:3 semi:4 full:1 sliding:4 keypoints:1 plug:1 long:2 retrieval:1 divided:1 controlled:1 regression:1 basic:1 vision:6 kernel:49 represent:2 whereas:4 want:1 addressed:1 else:1 publisher:1 sch:1 probably:1 strict:2 induced:2 tend:1 prague:1 practitioner:1 extracting:1 unitary:3 intermediate:1 iii:2 easy:1 split:1 fit:3 inner:11 translates:1 det:4 amnon:1 whether:1 pca:1 padding:1 ltd:1 suffer:1 algebraic:6 nine:1 useful:1 involve:1 amount:1 reduced:2 exist:1 notice:1 designer:1 sign:1 neuroscience:1 per:3 rb:1 edmonds:1 group:6 key:2 demonstrating:1 achieving:1 changing:1 padded:1 sum:7 run:1 angle:6 everywhere:1 letter:1 ameliorate:1 place:1 family:15 laid:1 sali:1 summarizes:1 submatrix:3 cca:1 bound:1 simplification:1 display:1 fold:8 encountered:1 strength:1 constraint:1 x2:4 sake:1 span:2 qb:9 separable:2 relatively:1 combination:10 kd:1 smaller:3 perm:1 making:1 invariant:4 taken:2 computationally:1 mutually:1 describing:1 fed:1 end:1 sending:3 operation:4 unreasonable:1 vidal:1 appropriate:2 alternative:1 schmidt:1 ho:1 jd:5 original:1 top:1 running:1 include:1 unifying:1 classical:1 tensor:12 noticed:1 pride:1 g0:1 added:1 exclusive:1 md:2 diagonal:5 gradient:1 subspace:10 mapped:2 ensuing:1 evenly:1 cauchy:2 discriminant:1 assuming:1 length:1 gx1:1 ratio:1 hebrew:1 difficult:1 expense:1 trace:6 negative:2 design:4 reliably:1 perform:1 upper:9 observation:3 commensurate:1 displayed:1 situation:1 extended:1 variability:1 y1:2 rn:10 sharp:1 jebara:1 intensity:1 introduced:1 pair:3 namely:1 required:3 engine:3 learned:2 czech:1 qa:6 below:5 pattern:1 articulation:2 summarize:1 built:2 max:1 analogue:4 power:4 brb:2 hr:2 innerproduct:1 representing:1 specializes:1 understanding:1 literature:1 geometric:1 graf:1 highlight:1 permutation:1 bear:1 prototypical:2 interesting:1 remarkable:1 localized:1 nels:1 degree:5 sufficient:1 article:1 principle:2 share:1 cd:10 row:4 aij:5 taking:1 face:1 emerge:2 sparse:2 van:1 dimension:8 xn:1 stand:2 tamir:1 quantum:1 collection:5 made:1 nonhomogenous:1 far:2 transaction:1 kullback:1 global:6 conclude:1 spectrum:1 table:4 nature:2 robust:1 caputo:1 complex:1 poly:2 european:2 domain:1 main:1 linearly:2 edition:2 body:2 x1:8 fig:2 referred:1 representative:1 fashion:1 sub:4 position:3 wish:2 exponential:1 jmlr:1 touched:1 down:7 unwieldy:1 rk:2 theorem:2 emphasized:1 jt:3 list:1 decay:1 svm:4 vapnik:1 effectively:1 lifting:6 demand:1 sophistication:1 lt:1 simply:1 appearance:3 forming:1 visual:3 ordered:1 partially:1 applies:1 springer:1 corresponds:1 wolf:1 acm:1 ma:1 symd:2 identity:4 viewed:2 rbf:7 fisher:1 naquet:1 change:3 hard:1 included:1 specifically:1 determined:1 loan:1 principal:5 called:1 total:2 invariance:12 formally:1 select:1 support:1 evaluate:1 rigidity:1 |
1,875 | 2,702 | Nonparametric Transforms of Graph Kernels
for Semi-Supervised Learning
Xiaojin Zhu?
?
Jaz Kandola?
School of Computer Science
Carnegie Mellon University
5000 Forbes Avenue
Pittsburgh, PA 15213 USA
Zoubin Ghahramani??
?
John Lafferty?
Gatsby Computational Neuroscience Unit
University College London
17 Queen Square
London WC1N 3AR UK
Abstract
We present an algorithm based on convex optimization for constructing
kernels for semi-supervised learning. The kernel matrices are derived
from the spectral decomposition of graph Laplacians, and combine labeled and unlabeled data in a systematic fashion. Unlike previous work
using diffusion kernels and Gaussian random field kernels, a nonparametric kernel approach is presented that incorporates order constraints
during optimization. This results in flexible kernels and avoids the need
to choose among different parametric forms. Our approach relies on
a quadratically constrained quadratic program (QCQP), and is computationally feasible for large datasets. We evaluate the kernels on real
datasets using support vector machines, with encouraging results.
1
Introduction
Semi-supervised learning has been the focus of considerable recent research. In this learning problem the data consist of a set of points, with some of the points labeled and the
remaining points unlabeled. The task is to use the unlabeled data to improve classification
performance. Semi-supervised methods have the potential to improve many real-world
problems, since unlabeled data are often far easier to obtain than labeled data.
Kernel-based methods are increasingly being used for data modeling and prediction because of their conceptual simplicity and good performance on many tasks. A promising
family of semi-supervised learning methods can be viewed as constructing kernels by transforming the spectrum of a ?local similarity? graph over labeled and unlabeled data. These
kernels, or regularizers, penalize functions that are not smooth over the graph [7]. Informally, a smooth eigenvector has the property that two elements of the vector have similar
values if there are many large weight paths between them on the graph. This results in the
desirable behavior of the labels varying smoothly over the graph, as sought by, e.g., spectral
clustering approaches [2], diffusion kernels [5], and the Gaussian random field approach
[9]. However, the modification to the spectrum, called a spectral transformation, is often
a function chosen from some parameterized family. As examples, for the diffusion kernel
the spectral transformation is an exponential function, and for the Gaussian field kernel the
transformation is a smoothed inverse function.
In using a parametric approach one faces the difficult problem of choosing an appropriate
family of spectral transformations. For many familes the number of degrees of freedom
in the parameterization may be insufficient to accurately model the data. In this paper
we propose an effective nonparametric method to find an optimal spectral transformation
using kernel alignment. The main advantage of using kernel alignment is that it gives
us a convex optimization problem, and does not suffer from poor convergence to local
minima. A key assumption of a spectral transformation is monotonicity, so that unsmooth
functions over the data graph are penalized more severly. We realize this property by
imposing order constraints. The optimization problem in general is solved using semidefinite programming (SDP) [1]; however, in our approach the problem can be formulated
in terms of quadratically constrained quadratic programming (QCQP), which can be solved
more efficiently than a general SDP.
This paper is structured as follows. In Section 2 we review some graph theoretic concepts
and relate them to the construction of kernels for semi-supervised learning. In Section 3
we introduce convex optimization via QCQP and relate it to the more familiar linear and
quadratic programming commonly used in machine learning. Section 4 poses the problem
of kernel based semi-supervised learning as a QCQP problem with order constraints. Experimental results using the proposed optimization framework are presented in Section 5.
The results indicate that the semi-supervised kernels constructed from the learned spectral
transformations perform well in practice.
2
Semi-supervised Kernels from Graph Spectra
We are given a labeled dataset consisting of input-output pairs {(x1 , y1 ), . . . , (xl , yl )} and
a (typically much larger) unlabeled dataset {xl+1 , . . . , xn } where x is in some general
input space and y is potentially from multiple classes. Our objective is to construct a kernel
that is appropriate for the classification task. Since our methods use both the labeled and
unlabeled data, we will refer to the resulting kernels as semi-supervised kernels. More
specifically, we restrict ourselves to the transductive setting where the unlabeled data also
serve as the test data. As such, we only need to find a good Gram matrix on the points
{x1 , . . . , xn }. For this approach to be effective such kernel matrices must also take into
account the distribution of unlabeled data, in order that the unlabeled data can aid in the
classification task. Once these kernel matrices have been constructed, they can be deployed
in standard kernel methods, for example support vector machines.
In this paper we motivate the construction of semi-supervised kernel matrices from a
graph theoretic perspective. A graph is constructed where the nodes are the data instances
{1, . . . , n} and an edge connects nodes i and j if a ?local similarity? measure between xi
and xj suggests they may have the same label. For example, the local similarity measure
can be the Euclidean distance between feature vectors if x ? Rm , and each node can connect to its k nearest neighbors with weight value equal to 1. The intuition underlying the
graph is that even if two nodes are not directly connected, they should be considered similar
as long as there are many paths between them. Several semi-supervised learning algorithms
have been proposed under the general graph theoretic theme, based on techniques such as
random walks [8], diffusion kernels [5], and Gaussian fields [9]. Many of these methods
can be unified into the regularization framework proposed by [7], which forms the basis of
this paper.
The graph can be represented by an n ? n weight matrix W = [wij ] where wij is the edge
weight between nodes i and j, with wij = 0 if there is no edge. We require the entries of W
to be non-negative, and assume that it forms a symmetric matrix; it is not necessary for W
itself to be positive semi-definite. In semi-supervised learning W is an essential quantity;
we assume it is provided by domain experts,
P and hence do not study its construction. Let
D be a diagonal matrix where dii =
j wij is the degree of node i. This allows us
to define the combinatorial graph Laplacian as L = D ? W (the normalized Laplacian
? = D?1/2 LD?1/2 can be used as well). We denote L?s eigensystem by {?i , ?i }, so
L
Pn
>
that L =
i=1 ?i ?i ?i where we assume the eigenvalues are sorted in non-decreasing
order. The matrix L has many interesting properties [3]; for instance, it is always positive
semi-definite, even if W is not. Perhaps the most important property of the Laplacian
related to semi-supervised learning is the following: a smaller P
eigenvalue ? corresponds
to a smoother eigenvector ? over the graph; that is, the value ij wij (?(i) ? ?(j))2 is
small. In a physical system the smoother eigenvectors correspond to the major vibration
modes. Assuming the graph structure is correct, from a regularization perspective we want
to encourage smooth functions, to reflect our belief that labels should vary slowly over the
graph. Specifically, [2] and [7] suggest a general principle for creating a semi-supervised
kernel K from the graph Laplacian L: transform the eigenvalues ? into r(?), where the
spectral transformation r is a non-negative and decreasing function1
K=
n
X
r(?i ) ?i ?>
i
(1)
i=1
Note that it may be that r reverses the order of the eigenvalues,
so that smooth ?i ?s have
P
larger eigenvalues in K. A ?soft labeling? function f =
cP
?
i i in a kernel machine has
a penalty term in the RKHS norm given by ?(||f ||2K ) = ?( c2i /r(?i )). Since r is decreasing, a greater penality is incurred for those terms of f corresponding to eigenfunctions
that are less smooth. In previous work r has often been chosen from a parametric family.
2
For example, the diffusion kernel [5] corresponds to r(?) = exp(? ?2 ?) and the Gaussian
1
field kernel [10] corresponds to r(?) = ?+
. Cross validation has been used to find the
hyperparameters ? or for these spectral transformations. Although the general principle
of equation (1) is appealing, it does not address question of which parametric family to use
for r. Moreover, the number of degrees of freedom (or the number of hyperparameters)
may not suit the task at hand, resulting in overly constrained kernels. The contribution of
the current paper is to address these limitations using a convex optimization approach by
imposing an ordering constraint on r but otherwise not assuming any parametric form for
the kernels.
3
Convex Optimization using QCQP
Let Ki = ?i ?>
matrices of the eigenvectors. The semii , i = 1 ? ? ? n be the outer productP
n
supervised kernel K is a linear combination K = i=1 ?i Ki , where ?i ? 0. We formulate
the problem of finding the spectral transformation as one that finds the interpolation coefficients {r(?i ) = ?i } by optimizing some convex objective function on K. To maintain the
positive semi-definiteness constraint on K, one in general needs to invoke SDPs [1]. Semidefinite optimization can be described as the problem of optimizing a linear function of a
symmetric matrix subject to linear equality constraints and the condition that the matrix be
positive semi-definite. The well-known linear programming problem can be generalized
to a semi-definite optimization by replacing the vector of variables with a symmetric matrix, and replacing the non-negativity constraints with a positive semi-definite constraints.
This generalization inherits several properties: it is convex, has a rich duality theory and
allows theoretically efficient solution algorithms based on iterating interior point methods
to either follow a central path or decrease a potential function. However, a limitation of
SDPs is their computational complexity [1], which has restricted their application to small
scale problems [6]. However, an important special case of SDPs are quadratically constrained quadratic programs (QCQP) which are computationally more efficient. Here both
the objective function and the constraints are quadratic as illustrated below,
minimize
subject to
1
1 >
x P0 x + q0> x + r0
2
1 >
x Pi x + qi> x + ri ? 0
2
Ax = b
(2)
i = 1???m
We use a slightly different notation where r is the inverse of that in [7].
(3)
(4)
n
n
where Pi ? S+
, i = 1, . . . , m, where S+
defines the set of square symmetric positive
semi-definite matrices. In a QCQP, we minimize a convex quadratic function over a feasible
region that is the intersection of ellipsoids. The number of iterations required to reach the
solution is comparable to the number required for linear programs, making the approach
feasible for large datasets. However, as observed in [1], not all SDPs can be relaxed to
QCQPs. For the semi-supervised kernel learning task presented here solving an SDP would
be computationally infeasible.
Recent work [4, 6] has proposed kernel target alignment that can be used not only to assess
the relationship between the feature spaces generated by two different kernels, but also to
assess the similarity between spaces induced by a kernel and that induced by the labels
themselves. Desirable properties of the alignment measure can be found in [4]. The crucial aspect of alignnement for our purposes is that its optimization can be formulated as a
QCQP. The objective function is the empirical kernel alignment score:
hKtr , T iF
? tr , T ) = p
A(K
hKtr , Ktr iF hT, T iF
(5)
where Ktr is the kernel matrix restricted to the training points,
P hM, N iF denotes the Frobenius product between two square matrices hM, N iF = ij mij nij = T r(M N > ), and T
is the target matrix on training data, with entry Tij set to +1 if yi = yj and ?1 otherwise.
Note for binary {+1, ?1} training labels y this is simply the rank one matrix T = yy> . K
is guaranteed to be positive semi-definite by constraining ?i ? 0. Previous work using kernel alignment did not take into account that the Ki ?s were derived from the graph Laplacian
with the goal of semi-supervised learning. As such, the ?i ?s can take arbitrary values and
there is no preference to penalize components that do not vary smoothly over the graph.
This can be rectified by requiring smoother eigenvectors to receive larger coefficients, as
shown in the next section.
4
Semi-Supervised Kernels with Order Constraints
As stated above, we would like to maintain a decreasing order on the spectral transformation ?i = r(?i ) to encourage smooth functions over the graph. This motivates the set of
order constraints
?i ? ?i+1 ,
i = 1???n ? 1
(6)
And we can specify the desired semi-supervised kernel as follows.
Definition 1 An order constrained semi-supervised kernel K is the solution to the following convex optimization problem:
maxK
subject to
? tr , T )
A(K
Pn
K = i=1 ?i Ki
?i ? 0
trace(K) = 1
?i ? ?i+1 ,
i = 1???n ? 1
(7)
(8)
(9)
(10)
(11)
where T is the training target matrix, Ki = ?i ?>
i and ?i ?s are the eigenvectors of the graph
Laplacian.
The formulation is an extension to [6] with order constraints, and with special components
Ki ?s from the graph Laplacian. Since ?i ? 0 and Ki ?s are outer products, K will automatically be positive semi-definite and hence a valid kernel matrix. The trace constraint is
needed to fix the scale invariance of kernel alignment. It is important to notice the order
constraints are convex, and as such the whole problem is convex. Let vec(A) be the column
vectorization of a matrix A. Defining M = vec(K1,tr ) ? ? ? vec(Km,tr ) , it is not hard to
show that the problem can then be expressed as
max?
subject to
vec(T )> M ?
||M ?|| ? 1
?i ? 0
?i ? ?i+1 ,
i = 1???n ? 1
(12)
(13)
(14)
(15)
The objective function is linear in ?, and there is a simple cone constraint, making it a
quadratically constrained quadratic program (QCQP).
An improvement of the above order constrained semi-supervised kernel can be obtained
by studying the Laplacian eigenvectors with zero eigenvalues. For a graph Laplacian there
will be k zero eigenvalues if the graph has k connected subgraphs. The k eigenvectors are
piecewise constant over individual subgraphs, and zero elsewhere. This is desirable when
k > 1, with the hope that subgraphs correspond to different classes. However if k = 1, the
graph is connected. The first eigenvector ?1 is a constant vector. The corresponding K1 is
a constant matrix, and acts as a bias term. In this situation we do not want to impose the
order constraint ?1 ? ?2 on the constant bias term. Instead we let ?1 vary freely during
optimization.
Definition 2 An improved order constrained semi-supervised kernel K is the solution to
the same problem in Definition 1, but the order constraints (11) apply only to non-constant
eigenvectors:
?i ? ?i+1 ,
i = 1 ? ? ? n ? 1, and ?i not constant
(16)
In practice we do not need all n eigenvectors of the graph Laplacian, or equivalently all n
Ki ?s. The first m < n eigenvectors with the smallest eigenvalues work well empirically.
Also note we could have used the fact that Ki ?s are from orthogonal eigenvectors ?i to
further simplify the expression. However we neglect this observation, making it easier to
incorporate other kernel components if necessary.
It is illustrative to compare and contrast the order constrained semi-supervised kernels to
other semi-supervised kernels with different spectral transformation. We call the original
kernel alignment solution in [6] a maximal-alignment kernel. It is the solution to Definition 1 without the order constraints (11). Because it does not have the additional constraints,
it maximizes kernel alignment among all spectral transformation. The hyperparameters ?
and of the Diffusion kernel and Gaussian fields kernel (described earlier) can be learned
by maximizing the alignment score also, although the optimization problem is not necessarily convex. These kernels use different information from the original Laplacian eigenvalues ?i . The maximal-alignment kernels ignore ?i altogether. The order constrained
semi-supervised kernels only use the order of ?i and ignore their actual values. The diffusion and Gaussian field kernels use the actual values. In terms of the degree of freedom in
choosing the spectral transformation ?i ?s, the maximal-alignment kernels are completely
free. The diffusion and Gaussian field kernels are restrictive since they have an implicit
parametric form and only one free parameter. The order constrained semi-supervised kernels incorporates desirable features from both approaches.
5
Experimental Results
We evaluate the order constrained kernels on seven datasets. baseball-hockey (1993 instances / 2 classes), pc-mac (1943/2) and religion-atheism (1427/2) are document categorization tasks taken from the 20-newsgroups dataset. The distance measure is the standard
cosine similarity between tf.idf vectors. one-two (2200/2), odd-even (4000/2) and ten
digits (4000/10) are handwritten digits recognition tasks. one-two is digits ?1? vs. ?2?;
odd-even is the artificial task of classifying odd ?1, 3, 5, 7, 9? vs. even ?0, 2, 4, 6, 8? digits,
such that each class has several well defined internal clusters; ten digits is 10-way classification. isolet (7797/26) is isolated spoken English alphabet recognition from the UCI
repository. For these datasets we use Euclidean distance on raw features. We use 10NN
unweighted graphs on all datasets except isolet which is 100NN. For all datasets, we use
the smallest m = 200 eigenvalue and eigenvector pairs from the graph Laplacian. These
values are set arbitrarily without optimizing and do not create a unfair advantage to the
proposed kernels. For each dataset we test on five different labeled set sizes. For a given
labeled set size, we perform 30 random trials in which a labeled set is randomly sampled
from the whole dataset. All classes must be present in the labeled set. The rest is used as
unlabeled (test) set in that trial. We compare 5 semi-supervised kernels (improved order
constrained kernel, order constrained kernel, Gaussian field kernel, diffusion kernel2 and
maximal-alignment kernel), and 3 standard supervised kernels (RBF (bandwidth learned
using 5-fold cross validation),linear and quadratic). We compute the spectral transformation for order constrained kernels and maximal-alignment kernels by solving the QCQP
using standard solvers (SeDuMi/YALMIP). To compute accuracy we use a standard SVM.
We choose the the bound on slack variables C with cross validation for all tasks and kernels. For multiclass classification we perform one-against-all and pick the class with the
largest margin.
The results3 are shown in Table 1, which has two rows for each cell: The upper row is
the average test set accuracy with one standard deviation; The lower row is the average
training set kernel alignment, and in parenthesis the average run time in seconds for SeDuMi/YALMIP on a 3GHz Linux computer. Each number is averaged over 30 random
trials. To assess the statistical significance of the results, we perform paired t-test on test
accuracy. We highlight the best accuracy in each row, and those that can not be determined
as different from the best, with paired t-test at significance level 0.05. The semi-supervised
kernels tend to outperform standard supervised kernels. The improved order constrained
kernels are consistently among the best. Figure 1 shows the spectral transformation ?i of
the semi-supervised kernels for different tasks. These are for the 30 trials with the largest
labeled set size in each task. The x-axis is in increasing order of ?i (the original eigenvalues
of the Laplacian). The mean (thick lines) and ?1 standard deviation (dotted lines) of only
the top 50 ?i ?s are plotted for clarity. The ?i values are scaled vertically for easy comparison among kernels. As expected the maximal-alignment kernels? spectral transformation is
zigzagged, diffusion and Gaussian field?s are very smooth, while order constrained kernels?
are in between. The order constrained kernels (green) have large ?1 because of the order
constraint. This seems to be disadvantageous ? the spectral transformation tries to balance
it out by increasing the value of other ?i ?s so that the constant K1 ?s relative influence is
smaller. On the other hand the improved order constrained kernels (black) allow ?1 to be
small. As a result the rest ?i ?s decay fast, which is desirable.
6
Conclusions
We have proposed and evaluated a novel approach for semi-supervised kernel construction
using convex optimization. The method incorporates order constraints, and the resulting
convex optimization problem can be solved efficiently using a QCQP. In this work the base
kernels were derived from the graph Laplacian, and no parametric form for the spectral
transformation was imposed, making the approach more general than previous approaches.
Experiments show that the method is both computationally feasible and results in improvements to classification performance when used with support vector machines.
2
The hyperparameters ? 2 and are learned with the fminbnd() function in Matlab to maximize
kernel alignment.
3
Results on baseball-hockey and odd-even are similar and omitted for space. Full results can be
found at http://www.cs.cmu.edu/?zhuxj/pub/ocssk.pdf
PC vs. MAC
Religion vs. Atheism
1
1
Improved order
Order
Max?align
Gaussian field
Diffusion
0.9
0.8
0.7
0.7
0.6
0.6
? scaled
? scaled
0.8
0.5
0.5
0.4
0.4
0.3
0.3
0.2
0.2
0.1
0
Improved order
Order
Max?align
Gaussian field
Diffusion
0.9
0.1
0
5
10
15
20
25
30
35
40
45
0
50
0
5
10
15
20
rank
Ten Digits (10 classes)
35
40
45
50
1
Improved order
Order
Max?align
Gaussian field
Diffusion
0.9
0.8
0.8
0.7
0.7
0.6
0.6
0.5
0.5
0.4
0.4
0.3
0.3
0.2
0.2
0.1
0.1
0
5
10
15
20
25
rank
30
35
40
45
Improved order
Order
Max?align
Gaussian field
Diffusion
0.9
? scaled
? scaled
30
ISOLET (26 classes)
1
0
25
rank
50
0
0
5
10
15
20
25
30
35
40
45
50
rank
Figure 1: Comparison of spectral transformation for the 5 semi-supervised kernels.
References
[1] S. Boyd and L. Vandenberge. Convex Optimization. Cambridge University Press, Cambridge
UK, 2004.
[2] O. Chapelle, J. Weston, and B. Sch?olkopf. Cluster kernels for semi-supervised learning. In
Advances in Neural Information Processing Systems, 15, volume 15, 2002.
[3] F. R. K. Chung. Spectral graph theory, Regional Conference Series in Mathematics, No. 92.
American Mathematical Society, 1997.
[4] N. Cristianini, J. Shawe-Taylor, A. Elisseeff, and J. Kandola. On kernel-target alignment. In
Advances in NIPS, 2001.
[5] R. I. Kondor and J. Lafferty. Diffusion kernels on graphs and other discrete input spaces. In
Proc. 19th International Conf. on Machine Learning, 2002.
[6] G. Lanckriet, N. Cristianini, P. Bartlett, L. E. Ghaoui, and M. Jordan. Learning the kernel matrix
with semidefinite programming. Journal of Machine Learning Research, 5:27?72, 2004.
[7] A. Smola and R. Kondor. Kernels and regularization on graphs. In Conference on Learning
Theory, COLT/KW, 2003.
[8] M. Szummer and T. Jaakkola. Partially labeled classification with Markov random walks. In
Advances in Neural Information Processing Systems, 14, volume 14, 2001.
[9] X. Zhu, Z. Ghahramani, and J. Lafferty. Semi-supervised learning using Gaussian fields and
harmonic functions. In ICML-03, 20th International Conference on Machine Learning, 2003.
[10] X. Zhu, J. Lafferty, and Z. Ghahramani. Semi-supervised learning: From Gaussian fields to
Gaussian processes. Technical Report CMU-CS-03-175, Carnegie Mellon University, 2003.
Training
set size
Improved
Order
pc-mac
10
87.0 ? 5.0
0.71 ( 1)
30
90.3 ? 1.3
0.68 ( 8)
50
91.3 ? 0.9
0.64 (31)
70
91.5 ? 0.6
0.63 (70)
90
91.5 ? 0.6
0.63 (108)
religion-atheism
10
72.8 ?11.2
0.50 ( 1)
30
84.2 ? 2.4
0.38 ( 8)
50
84.5 ? 2.3
0.31 (28)
70
85.7 ? 1.4
0.29 (55)
90
86.6 ? 1.3
0.27 (86)
one-two
10
96.2 ? 2.7
0.87 ( 2)
20
96.4 ? 2.8
0.87 ( 3)
30
98.2 ? 2.1
0.84 ( 8)
40
98.3 ? 1.9
0.84 (13)
50
98.4 ? 1.9
0.83 (31)
Ten digits (10 classes)
50
76.6 ? 4.3
0.47 (26)
100
84.8 ? 2.6
0.47 (124)
150
86.5 ? 1.7
0.48 (310)
200
88.1 ? 1.3
0.47 (708)
250
89.1 ? 1.1
0.47 (942)
isolet (26 classes)
50
56.0 ? 3.5
0.27 (26)
100
64.6 ? 2.1
0.26 (105)
150
67.6 ? 2.6
0.26 (249)
200
71.0 ? 1.8
0.26 (441)
250
71.8 ? 2.3
0.26 (709)
Order
semi-supervised kernels
Gaussian
Diffusion
Field
Max-align
RBF
standard kernels
Linear
Quadratic
84.9 ? 7.2
0.57 ( 1)
89.6 ? 2.3
0.49 ( 8)
90.5 ? 1.7
0.46 (31)
90.8 ? 1.3
0.46 (56)
91.3 ? 1.3
0.45 (98)
56.4 ? 6.2
0.32
76.4 ? 6.1
0.19
81.1 ? 4.6
0.16
84.6 ? 2.1
0.14
86.3 ? 2.3
0.13
57.8 ?11.5
0.35
79.6 ?11.2
0.23
87.5 ? 2.8
0.20
90.5 ? 1.2
0.19
91.3 ? 1.1
0.18
71.1 ? 9.7
0.90 ( 1)
85.4 ? 3.9
0.74 ( 6)
88.4 ? 2.1
0.68 (25)
89.6 ? 1.6
0.66 (59)
90.3 ? 1.0
0.65 (84)
51.6 ? 3.4
0.11
62.6 ? 9.6
0.03
67.8 ? 9.0
0.02
74.7 ? 7.4
0.01
79.0 ? 6.4
0.01
63.0 ? 5.1
0.30
71.8 ? 5.5
0.18
77.6 ? 4.8
0.14
80.2 ? 4.6
0.12
82.5 ? 4.2
0.11
62.3 ? 4.2
0.25
71.2 ? 5.3
0.13
75.7 ? 5.4
0.10
74.3 ? 8.7
0.08
79.1 ? 7.3
0.08
70.9 ?10.9
0.42 ( 1)
83.0 ? 2.9
0.31 ( 6)
83.5 ? 2.5
0.26 (23)
85.3 ? 1.6
0.25 (42)
86.4 ? 1.5
0.24 (92)
55.2 ? 5.8
0.31
71.2 ? 6.3
0.20
80.4 ? 4.1
0.17
83.0 ? 2.9
0.16
84.5 ? 2.1
0.15
60.9 ?10.7
0.31
80.3 ? 5.1
0.22
83.5 ? 2.7
0.20
85.4 ? 1.8
0.19
86.2 ? 1.6
0.18
60.7 ? 7.5
0.85 ( 1)
74.4 ? 5.4
0.60 ( 7)
77.4 ? 6.1
0.48 (27)
82.3 ? 3.0
0.43 (51)
82.8 ? 2.6
0.40 (85)
55.8 ? 5.8
0.13
63.4 ? 6.5
0.05
69.3 ? 6.5
0.04
73.1 ? 5.8
0.03
77.7 ? 5.1
0.02
60.1 ? 7.0
0.30
63.7 ? 8.3
0.18
69.4 ? 7.0
0.15
75.7 ? 6.0
0.13
74.6 ? 7.6
0.12
61.2 ? 4.8
0.26
70.1 ? 6.3
0.15
70.7 ? 8.5
0.11
71.0 ?10.0
0.10
70.0 ?11.5
0.09
90.6 ?14.0
0.66 ( 1)
93.9 ? 8.7
0.64 ( 4)
97.2 ? 2.5
0.61 ( 7)
96.5 ? 2.4
0.61 (15)
95.6 ? 9.0
0.60 (37)
58.2 ?17.6
0.43
87.0 ?16.0
0.38
98.1 ? 2.2
0.35
98.9 ? 1.8
0.36
99.4 ? 0.5
0.35
59.4 ?18.9
0.53
83.2 ?19.8
0.50
98.1 ? 2.7
0.47
99.1 ? 1.4
0.48
99.6 ? 0.3
0.46
85.4 ?11.5
0.95 ( 1)
94.5 ? 1.6
0.90 ( 3)
96.4 ? 2.1
0.86 ( 6)
96.3 ? 2.3
0.86 (11)
96.6 ? 2.3
0.84 (25)
78.7 ?14.3
0.38
90.4 ? 4.6
0.33
93.6 ? 3.1
0.30
94.0 ? 2.7
0.29
96.1 ? 2.4
0.28
85.1 ? 5.7
0.26
86.0 ? 9.4
0.22
89.6 ? 5.9
0.17
91.6 ? 6.3
0.18
93.0 ? 3.6
0.17
85.7 ? 4.8
0.30
90.9 ? 3.7
0.25
92.9 ? 2.8
0.24
94.9 ? 2.0
0.21
95.8 ? 2.3
0.20
71.5 ? 5.0
0.21 (26)
83.4 ? 2.6
0.17 (98)
86.4 ? 1.3
0.18 (255)
88.0 ? 1.3
0.16 (477)
89.3 ? 1.0
0.16 (873)
41.4 ? 6.8
0.15
63.7 ? 3.5
0.12
75.1 ? 3.0
0.11
80.4 ? 2.5
0.10
84.6 ? 1.4
0.10
49.8 ? 6.3
0.16
72.5 ? 3.3
0.13
80.4 ? 2.1
0.13
84.4 ? 1.6
0.11
87.2 ? 1.3
0.11
70.3 ? 5.2
0.51 (25)
80.7 ? 2.6
0.49 (100)
84.5 ? 1.9
0.50 (244)
86.0 ? 1.5
0.49 (523)
87.2 ? 1.3
0.49 (706)
57.0 ? 4.0
-0.62
69.4 ? 1.9
-0.64
75.2 ? 1.4
-0.66
78.3 ? 1.3
-0.65
80.4 ? 1.4
-0.65
50.2 ? 9.0
-0.50
56.0 ? 7.8
-0.52
56.2 ? 7.2
-0.53
60.8 ? 7.3
-0.54
61.3 ? 7.6
-0.54
66.3 ? 3.7
-0.25
77.2 ? 2.3
-0.29
81.4 ? 2.2
-0.31
84.3 ? 1.7
-0.33
85.7 ? 1.3
-0.33
42.0 ? 5.2
0.13 (25)
59.0 ? 3.6
0.10 (127)
65.2 ? 3.0
0.09 (280)
70.9 ? 2.3
0.08 (570)
73.6 ? 1.5
0.08 (836)
41.2 ? 2.9
0.03
58.5 ? 2.9
-0.02
65.4 ? 2.6
-0.05
70.6 ? 1.9
-0.07
73.7 ? 1.2
-0.07
29.0 ? 2.7
0.11
47.4 ? 2.7
0.08
57.2 ? 2.7
0.07
64.8 ? 2.1
0.06
69.8 ? 1.5
0.06
50.1 ? 3.7
0.31 (24)
63.2 ? 1.9
0.29 (102)
67.9 ? 2.5
0.27 (221)
72.3 ? 1.7
0.27 (423)
74.2 ? 1.5
0.27 (665)
28.7 ? 2.0
-0.89
46.3 ? 2.4
-0.90
57.6 ? 1.5
-0.90
63.9 ? 1.6
-0.91
68.8 ? 1.5
-0.91
30.0 ? 2.7
-0.80
46.6 ? 2.7
-0.82
57.3 ? 1.8
-0.83
64.2 ? 2.0
-0.83
69.5 ? 1.7
-0.84
23.7 ? 2.4
-0.65
42.0 ? 2.9
-0.69
53.8 ? 2.2
-0.70
60.5 ? 1.6
-0.72
66.2 ? 1.4
-0.72
Table 1: Accuracy, alignment scores, and run times on the datasets. The table compares 8
kernels. Each cell has two rows: The upper row is test set accuracy with standard error;
the lower row is training set alignment (SeDuMi/YALMIP run time in seconds is given in
parentheses). All numbers are averaged over 30 random trials. Accuracies in boldface are
the best as determined by a paired t-test at the 0.05 significance level.
| 2702 |@word trial:5 repository:1 kondor:2 norm:1 seems:1 km:1 decomposition:1 p0:1 elisseeff:1 pick:1 tr:4 ld:1 series:1 score:3 pub:1 rkhs:1 document:1 current:1 jaz:1 must:2 john:1 realize:1 v:4 parameterization:1 node:6 preference:1 five:1 mathematical:1 constructed:3 combine:1 introduce:1 theoretically:1 expected:1 behavior:1 themselves:1 sdp:3 decreasing:4 automatically:1 encouraging:1 actual:2 solver:1 increasing:2 provided:1 underlying:1 moreover:1 notation:1 maximizes:1 eigenvector:4 unified:1 finding:1 transformation:20 spoken:1 act:1 rm:1 scaled:5 uk:2 unit:1 positive:8 local:4 vertically:1 path:3 interpolation:1 black:1 suggests:1 averaged:2 yj:1 practice:2 definite:8 digit:7 empirical:1 boyd:1 zoubin:1 suggest:1 interior:1 unlabeled:11 influence:1 www:1 imposed:1 maximizing:1 convex:15 formulate:1 simplicity:1 subgraphs:3 isolet:4 construction:4 target:4 programming:5 lanckriet:1 pa:1 element:1 recognition:2 labeled:12 observed:1 solved:3 region:1 connected:3 ordering:1 decrease:1 intuition:1 transforming:1 complexity:1 cristianini:2 motivate:1 solving:2 serve:1 baseball:2 fminbnd:1 basis:1 completely:1 yalmip:3 represented:1 alphabet:1 fast:1 effective:2 london:2 artificial:1 labeling:1 choosing:2 larger:3 otherwise:2 transductive:1 transform:1 itself:1 advantage:2 eigenvalue:11 propose:1 product:2 maximal:6 uci:1 ktr:2 frobenius:1 olkopf:1 convergence:1 cluster:2 categorization:1 pose:1 ij:2 odd:4 nearest:1 school:1 c:2 indicate:1 revers:1 thick:1 correct:1 dii:1 require:1 fix:1 generalization:1 extension:1 considered:1 exp:1 major:1 sought:1 vary:3 smallest:2 omitted:1 purpose:1 proc:1 label:5 combinatorial:1 vibration:1 largest:2 tf:1 create:1 hope:1 gaussian:18 always:1 pn:2 varying:1 jaakkola:1 derived:3 focus:1 inherits:1 ax:1 improvement:2 consistently:1 rank:5 contrast:1 nn:2 typically:1 wij:5 among:4 flexible:1 classification:7 colt:1 constrained:19 special:2 field:17 construct:1 once:1 equal:1 kw:1 icml:1 report:1 piecewise:1 simplify:1 randomly:1 kandola:2 individual:1 familiar:1 consisting:1 ourselves:1 connects:1 maintain:2 suit:1 freedom:3 alignment:21 semidefinite:3 pc:3 regularizers:1 wc1n:1 edge:3 encourage:2 necessary:2 sedumi:3 orthogonal:1 euclidean:2 taylor:1 walk:2 desired:1 plotted:1 isolated:1 nij:1 instance:3 column:1 modeling:1 soft:1 earlier:1 ar:1 queen:1 mac:3 deviation:2 entry:2 connect:1 international:2 systematic:1 yl:1 invoke:1 kernel2:1 linux:1 reflect:1 central:1 choose:2 slowly:1 conf:1 creating:1 expert:1 chung:1 american:1 account:2 potential:2 coefficient:2 try:1 disadvantageous:1 forbes:1 contribution:1 square:3 minimize:2 ass:3 accuracy:7 efficiently:2 correspond:2 handwritten:1 sdps:4 raw:1 accurately:1 rectified:1 reach:1 definition:4 against:1 sampled:1 dataset:5 supervised:38 follow:1 specify:1 improved:9 formulation:1 evaluated:1 implicit:1 smola:1 hand:2 replacing:2 defines:1 mode:1 perhaps:1 usa:1 concept:1 normalized:1 requiring:1 regularization:3 hence:2 equality:1 symmetric:4 q0:1 illustrated:1 during:2 illustrative:1 cosine:1 generalized:1 eigensystem:1 pdf:1 theoretic:3 cp:1 harmonic:1 novel:1 physical:1 empirically:1 function1:1 volume:2 mellon:2 refer:1 cambridge:2 imposing:2 vec:4 mathematics:1 shawe:1 chapelle:1 similarity:5 base:1 align:5 recent:2 perspective:2 optimizing:3 binary:1 unsmooth:1 arbitrarily:1 yi:1 minimum:1 greater:1 relaxed:1 additional:1 impose:1 r0:1 freely:1 maximize:1 semi:44 smoother:3 multiple:1 desirable:5 full:1 smooth:7 technical:1 cross:3 long:1 paired:3 laplacian:14 qi:1 prediction:1 parenthesis:2 cmu:2 iteration:1 kernel:97 cell:2 penalize:2 receive:1 want:2 crucial:1 sch:1 rest:2 unlike:1 regional:1 eigenfunctions:1 subject:4 induced:2 tend:1 lafferty:4 incorporates:3 jordan:1 call:1 constraining:1 easy:1 newsgroups:1 xj:1 restrict:1 bandwidth:1 avenue:1 multiclass:1 expression:1 bartlett:1 penalty:1 suffer:1 matlab:1 tij:1 iterating:1 informally:1 eigenvectors:10 transforms:1 nonparametric:3 ten:4 http:1 outperform:1 notice:1 dotted:1 neuroscience:1 overly:1 yy:1 carnegie:2 discrete:1 key:1 clarity:1 diffusion:16 ht:1 graph:34 cone:1 run:3 inverse:2 parameterized:1 family:5 comparable:1 ki:9 bound:1 guaranteed:1 fold:1 quadratic:9 constraint:21 idf:1 ri:1 qcqp:11 aspect:1 structured:1 combination:1 poor:1 smaller:2 slightly:1 increasingly:1 appealing:1 modification:1 making:4 restricted:2 ghaoui:1 taken:1 computationally:4 equation:1 slack:1 needed:1 studying:1 apply:1 spectral:22 appropriate:2 altogether:1 original:3 denotes:1 remaining:1 clustering:1 top:1 neglect:1 restrictive:1 ghahramani:3 k1:3 society:1 objective:5 question:1 quantity:1 parametric:7 diagonal:1 distance:3 outer:2 seven:1 boldface:1 assuming:2 relationship:1 insufficient:1 ellipsoid:1 balance:1 equivalently:1 difficult:1 potentially:1 relate:2 trace:2 negative:2 stated:1 motivates:1 perform:4 upper:2 observation:1 datasets:8 markov:1 maxk:1 defining:1 situation:1 y1:1 smoothed:1 arbitrary:1 pair:2 required:2 quadratically:4 learned:4 nip:1 address:2 below:1 laplacians:1 program:4 max:6 green:1 belief:1 zhu:3 improve:2 axis:1 negativity:1 hm:2 xiaojin:1 review:1 relative:1 highlight:1 interesting:1 limitation:2 validation:3 incurred:1 degree:4 principle:2 classifying:1 pi:2 row:7 elsewhere:1 penalized:1 free:2 english:1 infeasible:1 bias:2 allow:1 neighbor:1 face:1 ghz:1 xn:2 world:1 avoids:1 gram:1 rich:1 valid:1 unweighted:1 commonly:1 far:1 ignore:2 monotonicity:1 conceptual:1 pittsburgh:1 severly:1 xi:1 spectrum:3 vectorization:1 hockey:2 table:3 promising:1 necessarily:1 constructing:2 domain:1 did:1 significance:3 main:1 whole:2 hyperparameters:4 atheism:3 x1:2 fashion:1 gatsby:1 deployed:1 aid:1 definiteness:1 theme:1 exponential:1 xl:2 unfair:1 decay:1 svm:1 consist:1 essential:1 margin:1 easier:2 smoothly:2 intersection:1 simply:1 expressed:1 religion:3 partially:1 mij:1 corresponds:3 relies:1 weston:1 viewed:1 formulated:2 sorted:1 goal:1 rbf:2 feasible:4 considerable:1 hard:1 specifically:2 except:1 determined:2 called:1 c2i:1 experimental:2 duality:1 invariance:1 college:1 internal:1 support:3 szummer:1 incorporate:1 evaluate:2 |
1,876 | 2,703 | The cerebellum chip:
an analog VLSI implementation of a
cerebellar model of classical conditioning
Constanze Hofst?tter, Manuel Gil, Kynan Eng,
Giacomo Indiveri, Matti Mintz, J?rg Kramer* and Paul F. M. J. Verschure
Institute of Neuroinformatics
University/ETH Zurich
CH-8057 Zurich, Switzerland
[email protected]
Abstract
We present a biophysically constrained cerebellar model of
classical conditioning, implemented using a neuromorphic analog
VLSI (aVLSI) chip. Like its biological counterpart, our cerebellar
model is able to control adaptive behavior by predicting the
precise timing of events. Here we describe the functionality of the
chip and present its learning performance, as evaluated in
simulated conditioning experiments at the circuit level and in
behavioral experiments using a mobile robot. We show that this
aVLSI model supports the acquisition and extinction of adaptively
timed conditioned responses under real-world conditions with
ultra-low power consumption.
I n tro d u cti o n
1
The association of two correlated stimuli, an initially neutral conditioned stimulus
(CS) which predicts a meaningful unconditioned stimulus (US), leading to the
acquisition of an adaptive conditioned response (CR), is one of the most essential
forms of learning. Pavlov introduced the classical conditioning paradigm in the
early 20th century to study associative learning (Pavlov 1927). In classical
conditioning training an animal is repeatedly exposed to a CS followed by a US
after a certain inter-stimulus interval (ISI). The animal learns to elicit a CR
matched to the ISI, reflecting its knowledge about an association between the CS,
US, and their temporal relationship. Our earlier software implementation of a
*
J?rg Kramer designed the cerebellum chip that was first tested at the 2002 Telluride
Neuromorphic Engineering Workshop. Tragically, he died soon afterwards while hiking
on Telescope Peak on 24 July, 2002.
biophysically constrained model of the cerebellar circuit underlying classical
conditioning (Verschure and Mintz 2001; Hofst?tter et al. 2002) provided an
explanation of this phenomenon by assuming a negative feedback loop between the
cerebellar cortex, deep nucleus and inferior olive. It could acquire and extinguish
correctly timed CRs over a range of ISIs in simulated classical conditioning
experiments, as well as in associative obstacle avoidance tasks using a mobile
robot. In this paper we present the analog VLSI (aVLSI) implementation of this
cerebellum model ? the cerebellum chip ? and the results of chip-level and
behavioral robot experiments.
2
T h e mo d el ci r cu i t a n d a VL S I i mp l eme n ta ti o n
Figure 1: Anatomy of the cerebellar model circuit (left) and the block diagram of
the corresponding chip (right).
The model (Figure 1) is based on the identified cerebellar pathways of CS, US and
CR (Kim and Thompson 1997) and includes four key hypotheses which were
implemented in the earlier software model (Hofst?tter et al. 2002):
1.
CS related parallel fiber (pf) and US related climbing fiber (cf) signals
converge at Purkinje cells (PU) in the cerebellum (Steinmetz et al. 1989). The
direction of the synaptic changes at the pf-PU-synapse depends on the temporal
coincidence of pf and cf activity. Long-term depression (LTD) is induced by pf
activity followed by cf activity within a certain time interval, while pf activity
alone induces long-term potentiation (LTP) (Hansel et al. 2001).
2.
A prolonged second messenger response to pf stimulation in the dendrites of
PU constitutes an eligibility trace from the CS pathway (Sutton and Barto
1990) that bridges the ISI (Fiala et al. 1996).
3.
A microcircuit (Ito 1984) comprising PU, deep nucleus (DN) and inferior olive
(IO) forms a negative feedback loop. Shunting inhibition of IO by DN blocks
the reinforcement pathway (Thompson et al. 1998), thus controlling the
induction of LTD and LTP at the pf-PU-synapse.
4.
DN activity triggers behavioral CRs (McCormick and Thompson 1984). The
inhibitory PU controls DN activity by a mechanism called rebound excitation
(Hesslow 1994): When DN cells are disinhibited from PU input, their
membrane potential slowly repolarises and spikes are emitted if a certain
threshold is reached. Thereby, the correct timing of CRs results from the
adaptation of a pause in PU spiking following the CS.
In summary, in the model the expression of a CR is triggered by DN rebound
excitation upon release from PU inhibition. The precise timing of a CR is
dependent on the duration of an acquired pause in PU spiking following a CS. The
PU response is regulated by LTD and LTP at the pf-PU-synapse under the control
of a negative feedback loop comprising DN, PU and IO.
We implemented an analog VLSI version of the cerebellar model using a standard
1.6?m CMOS technology, and occupying an area of approximately 0.25 mm2. A
block diagram of the hardware model is shown in Figure 1. The CS block receives
the conditioned stimulus and generates two signals: an analog long-lasting, slowly
decaying trace (cs_out) and an equally long binary pulse (cs_wind). Similarly, the
US block receives an unconditioned stimulus and generates a fast pulse (us_out).
The two pulses cs_wind and us_out are sent to the LT-ISI block that is responsible
for perfoming LTP and LTD, upregulating or downregulating the synaptic weight
signal w. This signal determines the gain by which the cs_out trace is multiplied in
the MU block. The output of the multiplier MU is sent on to the PU block, together
with the us_out signal. It is a linear integrate-and-fire neuron (the axon-hillock
circuit) connected to a constant current source that produces regular spontaneous
activity. The current source is gated by the digital cf_wind signal, such that the
spontaneous activity is shut off for the duration of the cs_out trace.
The chip allowed one of three learning rules to be connected. Experiments showed
that an ISI-dependent learning rule with short ISIs resulting in the strongest LTD
was the most useful (Kramer and Hofst?tter 2002). Two elements were added to
adapt the model circuit for real-world robot experiments. Firstly, to prevent the
expression of a CR after a US had already been triggered, an inhibitory connection
from IO to CRpathway was added. Secondly, the transduction delay (TD) from the
aVLSI circuit to any effectors (e.g. motor controls of a robot) had to be taken into
account, which was done by adding a delay from DN to IO of 500ms.
The chip?s power consumption is conservatively estimated at around 100 W
(excluding off-chip interfacing), based on measurements from similar integrateand-fire neuron circuits (Indiveri 2003). This figure is an order of magnitude lower
than what could be achieved using conventional microcontrollers (typically 1-10
mW), and could be improved further by optimising the circuit design.
3
S i mu l a ted co n d i ti o n i n g ex p eri men ts
The aim of the ?in vitro? simulated conditioning experiments was to understand the
learning performance of the chip. To obtain a meaningful evaluation of the
performance of the learning system for both the simulated conditioning
experiments and the robot experiments, the measure of effective CRs was used. In
acquisition experiments CS-US pairs are presented with a fixed ISI. Whenever a
CR occurs that precedes the US, the US signal is not propagated to PU due to the
inhibitory connection from DN to IO. Thus in the context of acquisition
experiments a CR is defined as effective if it prevents the occurrence of a US spike
at PU. In contrast, in robot experiments an effective CR is defined at the
behavioral level, including only CRs that prevent the US from occurring.
Figure 2: Learning related response changes in the cerebellar aVLSI chip. The most
relevant neural responses to a CS-US pair (ISI of 3s, ITI of 12s) are presented for a
trial before (naive) significant learning occurred and when a correctly timed CR is
expressed (trained). US-related pf and CS/CR-related cf signals are indicated by
vertical lines passing through the subplots. A CS-related pf-signal evokes a
prolonged response in the pf-PU-synapse, the CS-trace (Trace subplot). While an
active CS-trace is present, an inhibitory element (I) is active which inactivates an
element representing the spontaneous activity of PU (Hofst?tter et al. 2002). (A)
The US-related cf input occurs while there is an active CS-trace (Trace subplot), in
this case following the CS with an ISI of 3s. LTD predominates over LTP under
these conditions (Weight subplot). Because the PU membrane potential (PU)
remains above spiking threshold, PU is active and supplies constant inhibition to
DN (DN) while in the CS-mode. Thus, DN cannot repolarize and remains inactive
so that no CR is triggered. (B) Later in the experiment, the synaptic weight of the
pf-PU-synapse (Weight) has been reduced due to previous LTD. As a result,
following a CS-related pf input, the PU potential (PU subplot) falls below the
spiking threshold, which leads to a pause in PU spiking. The DN membrane
potential repolarises, so that rebound spikes are emitted (DN subplot). This
rebound excitation triggers a CR. DN inhibition of IO prevents US related cfactivity. Thus, although a US signal is still presented to the circuit, the reinforcing
US pathway is blocked. These conditions induce only LTP, raising the synaptic
weight of the pf-PU-synapse (Weight subplot).
The results we obtained were broadly consistent with those reported in the
biological literature (Ito 1984; Kim and Thompson 1997). The correct operation of
the circuit can be seen in the cell traces illustrating the properties of the aVLSI
circuit components before significant learning (Figure 2 A), and after a CR is
expressed (Figure 2B). Long-term acquisition experiments (25 blocks of 10 trials
each over 50 minutes) showed that chip functions remained stable over a long time
period. In each trial the CS was followed by a US with a fixed ISI of 3s; the inter
trial interval (ITI) was 12s. The number of effective CRs shows an initial fast
learning phase followed by a stable phase with higher percentages of effective CRs
(Figure 3B). In the stable phase the percentage of effective CRs per block
fluctuates around 80-90%. There are fluctuations of up to 500ms in the CR latency
caused by the interaction of LTD and LTP in the stable phase, but the average CR
latency remains fairly constant.
Figure 4 shows the average of five acquisition experiments (5 blocks of 10 trials
per experiment) for ISIs of 2.5s, 3s and 3.5s. The curves are similar in shape to the
ones in the long-term experiment. The CR latency quickly adjusts to match the ISI
and remains stable thereafter (Figure 4A). The effect of the ISI-dependent learning
rule can be seen in two ways: firstly, the shorter the ISI, the faster the stable phase
is reached, denoting faster learning. Secondly, the shorter the ISI, the better the
performance in terms of percentage of effective CRs (Figure 4B). The parameters
of the chip were tuned to optimally encode short ISIs in the range of 1.75s to 4.5s.
Separate experiments showed that the chip could also adapt rapidly to changes in
the ISI within this range after initial learning.
(Error bar = 1 std. dev.)
Figure 3: Long-term changes in CR latency (A) and % effective CRs (B) per block
of 10 CSs during acquisition. Experiment length = 50min., ISI = 3s, ITI = 12s.
(Error bar = 1 std. dev.)
Figure 4: Average of five acquisition experiments per block of 10 CSs for ISIs of
2.5s ( ), 3s (*) and 3.5s ( ). (A) Avg. CR latency. (B) Avg. % effective CRs.
4
Ro b o t a s s o ci a ti v e l ea rn i n g ex p eri men t s
The ?in vivo? learning capability of the chip was evaluated by interfacing it to a
robot and observing its behavior in an unsupervised obstacle avoidance task.
Experiments were performed using a Khepera microrobot (K-team, Lausanne,
Switzerland, Figure 5A) in a circular arena with striped walls (Figure 5C). The
robot was equipped with 6 proximal infra-red (IR) sensors (Figure 5B). Activation
of these sensors (US) due to a collision triggered a turn of ~110? in the opposite
direction (UR). A line camera (64 pixels x 256 gray-levels) constituted the distal
sensor, with detection of a certain spatial frequency (~0.14 periods/degree)
signalling the CS. Visual CSs and collision USs were conveyed to CSpathway and
USpathway on the chip. The activation of CRpathway triggered a motor CR: a 1s
long regression followed by a turn of ~180?. Communication between the chip and
the robot was performed using Matlab on a PC. The control program could be
downloaded to the robot's processor, allowing the robot to act fully autonomously.
In each experiment, the robot was placed in the circular arena exploring its
environment with a constant speed of ~4 cm/s. A spatial frequency CS was
detected at some distance when the robot approached the wall, followed by a
collision with the wall, stimulating the IR sensors and thus triggering a US.
Consequently the CS was correlated with the US, predicting it. The ISIs of these
stimuli were variable, due to noise in sensor sampling, and variations in the angle
at which the robot approached the wall.
Figure 5: (A) Khepera microrobot with aVLSI chip mounted on top. (B) Only the
forward sensors were used during the experiments. (C) The environment: a 60cm
diameter circular arena surrounded by a 15cm high wall. A pattern of vertical,
equally sized black and white bars was placed on the wall.
Associative learning mediated by the cerebellum chip significantly altered the
robot's behavior in the obstacle avoidance task (Figure 6) over the course of each
experiment. In the initial learning phase, the behavior was UR driven: the robot
drove forwards until it collided with the wall, only then performing a turn (Figure
6A1). In the trained phase, the robot usually turned just before it collided with the
wall (Figure 6A2), reducing the number of collisions. The positions of the robot
when a CS, US or CR event occurred in these two phases are shown in Figure 6B1
and B2. The CRs were not expressed immediately after the CSs, but rather with a
CR latency adjusted to just prevent collisions (USs). Not all USs were avoided in
the trained phase due to some excessively short ISIs (Figure 7) and normal
extinction processes over many unreinforced trials. After the learning phase the
percentage of effective CRs fluctuated between 70% and 100% (Figure 7).
Figure 6: Learning performance of the robot. (Top row) Trajectories of the robot.
The white circle with the black dot in the center indicates the beginning of
trajectories. (Bottom row) The same periods of the experiment examined at the
circuit level: = CS, * = US, = CR. (A1, B1) Beginning of the experiment (CS
3-15). (A2, B2) Later in the experiment (CS 32-44).
Figure 7: Trends in learning behavior (average of 5 experiments, 25 min. each). 90
CSs were presented in each experiment. Error bars indicate one standard deviation.
(A) Average percentage of effective CRs over 9 blocks of 10 CSs. (B) Number of
CS occurrences ( ), US occurrences (*) and CR occurrences ( ).
5
Di s cu s s i o n
We have presented one of the first examples of a biologically constrained model of
learning implemented in hardware. Our aVLSI cerebellum chip supports the
acquisition and extinction of adaptively timed responses under noisy, real world
conditions. These results provide further evidence for the role of the cerebellar
circuit embedded in a synaptic feedback loop in the learning of adaptive behavior,
and pave the way for the creation of artefacts with embedded ultra low-power
learning capabilities.
6
Ref eren ces
Fiala, J. C., Grossberg, S. and Bullock, D. (1996). Metabotropic glutamate receptor
activation in cerebellar Purkinje cells as substrate for adaptive timing of the classical
conditioned eye-blink response. Journal of Neuroscience 16: 3760-3774.
Hansel, C., Linden, D. J. and D'Angelo, E. (2001). Beyond parallel fiber LTD, the diversity
of synaptic and nonsynaptic plasticity in the cerebellum. Nature Neuroscience 4: 467-475.
Hesslow, G. (1994). Inhibition of classical conditioned eyeblink response by stimulation of
the cerebellar cortex in decerebrate cat. Journal of Physiology 476: 245-256.
Hofst?tter, C., Mintz, M. and Verschure, P. F. M. J. (2002). The cerebellum in action: a
simulation and robotics study. European Journal of Neuroscience 16: 1361-1376.
Indiveri, G. (2003). A low-power adaptive integrate-and-fire neuron circuit. IEEE
International Symposium on Circuits and Systems, Bangkok, Thailand, 4: 820-823.
Ito, M. (1984). The modifiable neuronal network of the cerebellum. Japanese Journal of
Physiology 5: 781-792.
Kim, J. J. and Thompson, R. F. (1997). Cerebellar circuits and synaptic mechanisms
involved in classical eyeblink conditioning. Trends in the Neurosciences 20(4): 177-181.
Kim, J. J. and Thompson, R. F. (1997). Cerebellar circuits and synaptic mechanisms
involved in classical eyeblink conditioning. Trend. Neurosci. 20: 177-181.
Kramer, J. and Hofst?tter, C. (2002). An aVLSI model of cerebellar mediated associative
learning. Telluride Workshop, CO, USA.
McCormick, D. A. and Thompson, R. F. (1984). Neuronal response of the rabbit cerebellum
during acquisition and performance of a classical conditioned nictitating membrane-eyelid
response. J. Neurosci. 4: 2811-2822.
Pavlov, I. P. (1927). Conditioned Reflexes, Oxford University Press.
Steinmetz, J. E., Lavond, D. G. and Thompson, R. F. (1989). Classical conditioning in
rabbits using pontine nucleus stimulation as a conditioned stimulus and inferior olive
stimulation as an unconditioned stimulus. Synapse 3: 225-233.
Sutton, R. S. and Barto, A. G. (1990). Time derivate models of Pavlovian Reinforcement
Learning and Computational Neuroscience: Foundations of Adaptive Networks., MIT press:
chapter 12, 497-537.
Thompson, R. F., Thompson, J. K., Kim, J. J. and Shinkman, P. G. (1998). The nature of
reinforcement in cerebellar learning. Neurobiology of Learning and Memory 70: 150-176.
Verschure, P. F. M. J. and Mintz, M. (2001). A real-time model of the cerebellar circuitry
underlying classical conditioning: A combined simulation and robotics study.
Neurocomputing 38-40: 1019-1024.
| 2703 |@word trial:6 cu:2 version:1 illustrating:1 extinction:3 pulse:3 simulation:2 eng:1 thereby:1 initial:3 denoting:1 tuned:1 current:2 manuel:1 activation:3 olive:3 plasticity:1 shape:1 motor:2 designed:1 alone:1 shut:1 signalling:1 beginning:2 short:3 firstly:2 five:2 dn:15 supply:1 symposium:1 pathway:4 behavioral:4 acquired:1 inter:2 behavior:6 isi:22 td:1 prolonged:2 pf:14 equipped:1 provided:1 matched:1 underlying:2 circuit:17 what:1 cm:3 temporal:2 ti:3 act:1 ro:1 control:5 before:3 engineering:1 timing:4 died:1 nictitating:1 io:7 sutton:2 receptor:1 oxford:1 fluctuation:1 approximately:1 black:2 inactivates:1 examined:1 khepera:2 pavlov:3 co:2 lausanne:1 range:3 grossberg:1 responsible:1 camera:1 block:14 area:1 elicit:1 eth:1 significantly:1 physiology:2 induce:1 regular:1 cannot:1 context:1 conventional:1 center:1 duration:2 thompson:10 rabbit:2 immediately:1 rule:3 avoidance:3 adjusts:1 century:1 variation:1 cs:6 controlling:1 trigger:2 spontaneous:3 drove:1 substrate:1 us:3 hypothesis:1 element:3 trend:3 std:2 predicts:1 hesslow:2 bottom:1 role:1 coincidence:1 connected:2 autonomously:1 environment:2 mu:3 trained:3 exposed:1 creation:1 upon:1 chip:21 cat:1 fiber:3 chapter:1 fast:2 describe:1 effective:11 precedes:1 detected:1 approached:2 neuroinformatics:1 fluctuates:1 noisy:1 unconditioned:3 associative:4 triggered:5 interaction:1 adaptation:1 relevant:1 loop:4 turned:1 rapidly:1 produce:1 cmos:1 avlsi:9 implemented:4 c:28 indicate:1 switzerland:2 direction:2 anatomy:1 artefact:1 correct:2 functionality:1 potentiation:1 integrateand:1 wall:8 ultra:2 biological:2 secondly:2 adjusted:1 exploring:1 around:2 normal:1 mo:1 circuitry:1 early:1 a2:2 angelo:1 hansel:2 bridge:1 occupying:1 mit:1 interfacing:2 sensor:6 aim:1 rather:1 cr:38 mobile:2 barto:2 encode:1 release:1 indiveri:3 indicates:1 contrast:1 kim:5 dependent:3 el:1 vl:1 typically:1 initially:1 vlsi:4 collided:2 comprising:2 pixel:1 animal:2 constrained:3 spatial:2 fairly:1 ted:1 sampling:1 mm2:1 optimising:1 unsupervised:1 constitutes:1 rebound:4 infra:1 stimulus:9 steinmetz:2 neurocomputing:1 mintz:4 phase:10 fire:3 detection:1 circular:3 evaluation:1 arena:3 pc:1 shorter:2 timed:4 circle:1 effector:1 earlier:2 obstacle:3 purkinje:2 dev:2 neuromorphic:2 deviation:1 neutral:1 delay:2 optimally:1 hillock:1 reported:1 proximal:1 giacomo:1 combined:1 adaptively:2 peak:1 international:1 off:2 predominates:1 together:1 quickly:1 slowly:2 leading:1 account:1 potential:4 fiala:2 diversity:1 b2:2 includes:1 mp:1 caused:1 depends:1 later:2 performed:2 observing:1 reached:2 red:1 decaying:1 parallel:2 capability:2 vivo:1 ir:2 climbing:1 blink:1 biophysically:2 trajectory:2 processor:1 strongest:1 phys:1 messenger:1 whenever:1 synaptic:8 acquisition:10 frequency:2 involved:2 di:1 propagated:1 gain:1 knowledge:1 ea:1 reflecting:1 ta:1 higher:1 response:12 improved:1 synapse:7 evaluated:2 microcircuit:1 done:1 unreinforced:1 just:2 until:1 receives:2 mode:1 gray:1 indicated:1 usa:1 effect:1 excessively:1 multiplier:1 counterpart:1 metabotropic:1 white:2 distal:1 cerebellum:11 during:3 eligibility:1 inferior:3 excitation:3 m:2 ini:1 tro:1 stimulation:4 spiking:5 vitro:1 conditioning:13 analog:5 association:2 he:1 occurred:2 measurement:1 significant:2 blocked:1 similarly:1 had:2 dot:1 decerebrate:1 robot:21 stable:6 cortex:2 inhibition:5 pu:26 showed:3 driven:1 certain:4 binary:1 seen:2 subplot:6 converge:1 paradigm:1 period:3 july:1 signal:10 afterwards:1 match:1 adapt:2 faster:2 long:9 shunting:1 equally:2 a1:2 regression:1 cerebellar:17 achieved:1 cell:4 robotics:2 derivate:1 interval:3 diagram:2 source:2 induced:1 ltp:7 sent:2 emitted:2 mw:1 subplots:1 identified:1 opposite:1 triggering:1 inactive:1 expression:2 ltd:9 reinforcing:1 passing:1 repeatedly:1 depression:1 deep:2 matlab:1 useful:1 latency:6 collision:5 action:1 thailand:1 induces:1 hardware:2 telescope:1 reduced:1 diameter:1 percentage:5 inhibitory:4 gil:1 estimated:1 neuroscience:5 correctly:2 per:4 modifiable:1 broadly:1 key:1 four:1 thereafter:1 threshold:3 prevent:3 ce:1 eme:1 angle:1 evokes:1 followed:6 activity:9 hiking:1 striped:1 software:2 generates:2 speed:1 min:2 performing:1 pavlovian:1 membrane:4 ur:2 bullock:1 biologically:1 lasting:1 taken:1 zurich:2 remains:4 turn:3 mechanism:3 operation:1 multiplied:1 occurrence:4 top:2 cf:5 eri:2 classical:13 added:2 already:1 spike:3 occurs:2 pave:1 regulated:1 distance:1 separate:1 simulated:4 consumption:2 induction:1 assuming:1 length:1 tter:7 relationship:1 tragically:1 acquire:1 trace:10 negative:3 implementation:3 design:1 gated:1 mccormick:2 allowing:1 vertical:2 neuron:3 iti:3 t:1 neurobiology:1 excluding:1 precise:2 team:1 communication:1 rn:1 introduced:1 pair:2 connection:2 raising:1 able:1 bar:4 beyond:1 below:1 pattern:1 usually:1 program:1 including:1 memory:1 explanation:1 power:4 event:2 predicting:2 glutamate:1 pause:3 representing:1 altered:1 technology:1 eye:1 mediated:2 naive:1 literature:1 embedded:2 fully:1 men:2 mounted:1 digital:1 foundation:1 nucleus:3 integrate:2 degree:1 conveyed:1 downloaded:1 consistent:1 surrounded:1 row:2 course:1 summary:1 placed:2 soon:1 verschure:4 understand:1 institute:1 fall:1 eyelid:1 feedback:4 curve:1 world:3 conservatively:1 forward:2 adaptive:6 reinforcement:3 avg:2 avoided:1 active:4 b1:2 pontine:1 nature:2 matti:1 correlated:2 dendrite:1 european:1 japanese:1 constituted:1 neurosci:2 noise:1 paul:1 allowed:1 ref:1 neuronal:2 transduction:1 eyeblink:3 axon:1 position:1 bangkok:1 ito:3 learns:1 minute:1 remained:1 linden:1 evidence:1 essential:1 workshop:2 adding:1 extinguish:1 ci:2 magnitude:1 conditioned:9 occurring:1 rg:2 lt:1 visual:1 repolarize:1 prevents:2 expressed:3 reflex:1 ch:2 determines:1 fluctuated:1 stimulating:1 cti:1 telluride:2 kramer:4 sized:1 consequently:1 change:4 reducing:1 called:1 meaningful:2 support:2 ethz:1 tested:1 phenomenon:1 ex:2 |
1,877 | 2,704 | VDCBPI: an Approximate Scalable Algorithm
for Large POMDPs
Pascal Poupart
Department of Computer Science
University of Toronto
Toronto, ON M5S 3H5
[email protected]
Craig Boutilier
Department of Computer Science
University of Toronto
Toronto, ON M5S 3H5
[email protected]
Abstract
Existing algorithms for discrete partially observable Markov decision
processes can at best solve problems of a few thousand states due to
two important sources of intractability: the curse of dimensionality and
the policy space complexity. This paper describes a new algorithm
(VDCBPI) that mitigates both sources of intractability by combining the
Value Directed Compression (VDC) technique [13] with Bounded Policy Iteration (BPI) [14]. The scalability of VDCBPI is demonstrated on
synthetic network management problems with up to 33 million states.
1 Introduction
Partially observable Markov decision processes (POMDPs) provide a natural and expressive framework for decision making, but their use in practice has been limited by the lack
of scalable solution algorithms. Two important sources of intractability plague discrete
model-based POMDPs: high dimensionality of belief space, and the complexity of policy
or value function (VF) space. Classic solution algorithms [4, 10, 7], for example, compute
value functions represented by exponentially many value vectors, each of exponential size.
As a result, they can only solve POMDPs with on the order of 100 states. Consequently,
much research has been devoted to mitigating these two sources of intractability.
The complexity of policy/VF space has been addressed by observing that there are often
very good policies whose value functions are representable by a small number of vectors.
Various algorithms such as approximate vector pruning [9], point-based value iteration
(PBVI) [12, 16], bounded policy iteration (BPI) [14], gradient ascent (GA) [11, 1] and
stochastic local search (SLS) [3] exploit this fact to produce (often near-optimal) policies
of low complexity (i.e., few vectors) allowing larger POMDPs to be solved. Still these
scale to problems of only roughly 1000 states, since each value vector may still have exponential dimensionality. Conversely, it has been observed that belief states often carry
more information than necessary. Hence, one can often reduce vector dimensionality by
using compact representations such as decision trees (DTs) [2], algebraic decision diagrams (ADDs) [8, 9], or linear combinations of small basis functions (LCBFs) [6], or by
indirectly compressing the belief space into a small subspace by a value-directed compression (VDC) [14] or exponential PCA [15]. Once compressed, classic solution methods can
be used. However, since none of these approaches address the exponential complexity of
policy/VF space, they can only solve slightly larger POMDPs (up to 8250 states [15]).
Scalable POMDP algorithms can only be realized when both sources of intractability are
tackled simultaneously. While Hansen and Feng [9] implemented such an algorithm by
combining approximate state abstraction with approximate vector pruning, they didn?t
demonstrate the scalability of the approach on large problems. In this paper, we describe
how to combine value directed compression (VDC) with bounded policy iteration (BPI)
and demonstrate the scalability of the resulting algorithm (VDCBPI) on synthetic network
management problems of up to 33 million states. Among the techniques that deal with the
curse of dimensionality, VDC offers the advantage that the compressed POMDP can be directly fed into existing POMDP algorithms with no (or only slight) adjustments. This is not
the case for exponential-PCA, nor compact representations (DTs, ADDs, LCBFs). Among
algorithms that mitigate policy space complexity, BPI distinguishes itself by its ability to
avoid local optima (cf. GA), its efficiency (cf. SLS) and the fact that belief state monitoring
is not required (cf. PBVI, approximate vector pruning). Beyond the combination of VDC
with BPI, we offer two other contributions. We propose a new simple heuristic to compute
good lossy value directed compressions. We also augment BPI with the ability to bias its
policy search to reachable belief states. As a result, BPI can often find a much smaller
policy of similar quality for a given initial belief state.
2 POMDP Background
A POMDP is defined by: states S; actions A; observations Z; transition function T , where
T (s, a, s0 ) denotes Pr(s0 |s, a); observation function Z, where Z(s, z) is the probability
Pr(z|s, a) of observation z in state s after executing a; and reward function R, where
R(s, a) is the immediate reward associated with s when executing a. We assume discrete
state, action and observation sets and focus on discounted, infinite horizon POMDPs with
discount factor 0 ? ? < 1.
Policies and value functions for POMDPs are typically defined over belief space B, where
a belief state b is a distribution over S capturing an agent?s knowledge about the current
state of the world. Belief state b can be updated in response to a specific action-observation
pair ha, zi using Bayes rule. We denote the (unnormalized) belief update mapping by T a,z ,
where Tija,z = Pr(sj |a, si ) Pr(z|sj ). A factored POMDP, with exponentially many states,
thus gives rise to a belief space of exponential dimensionality.
Policies represented by finite state controllers (FSCs) are defined by a (possibly cyclic) directed graph ? = hN , Ei, where nodes n ? N correspond to stochastic action choices and
edges e ? E to stochastic transitions. An FSC can be viewed as a policy ? = h?, ?i, where
action strategy ? associates each node n with a distribution over actions ?(n) = Pr(a|n),
and observation strategy ? associates each node n and observation z with a distribution
over successor nodes ?(n, z) = Pr(n0 |n, z) (corresponding to the edge from n labeled
with z). The value function V ? of FSC ? is given by:
X
X
X
?
0
0
0
?
0
0
V (n, s) =
Pr(a|n)R(s, a) + ?
a
Pr(s |s, a) Pr(z|s , a)
Pr(n |n, z)V (n , s ) (1)
n0
z
The value V (n, b) of each node n is thus linear w.r.t the belief state; hence the value
function of the controller is piecewise-linear and convex. The optimal value function V ?
often has a large (if not infinite) number of vectors, each corresponding to a different node.
The optimal value function V ? satisfies Bellman?s equation:
X
?
a
V (b) = max R(b, a) + ?
Pr(z|b, a)V (bz )
a
z
(2)
max
s.t.
V (n,
P
Ps) + ?
0
|s, a) Pr(z|s0 , a) Pr(a, n0 |n, z)V (n0 , s0 )], ?s
P a [Pr(a|n)R(s, a)P+ ? s0 ,z Pr(s
0
Pr(a|n) = 1;
Pr(a, n |n, z) = Pr(a|n), ?a
a
n0
Pr(a|n) ? 0, ?a;
Pr(a, n0 |n, z) ? 0, ?a, z
Table 1: LP to uniformly improve the value function of a node.
max
s.t.
P
o(s, n)s,n
s,n
V (n,
s)
P + s,n ?
P
0
|s, a) Pr(z|s0 , a) Pr(a, n0 |n, z)V (n0 , s0 )], ?s
P a [Pr(a|n)R(s, a)P+ ? s0 ,z Pr(s
0
Pr(a|n) = 1;
Pr(a, n |n, z) = Pr(a|n), ?a
a
n0
Pr(a|n) ? 0, ?a;
Pr(a, n0 |n, z) ? 0, ?a, z
Table 2: LP to improve the value function of a node in a non-uniform way according to the
steady state occupancy o(s, n).
3 Bounded Policy Iteration
We briefly review the bounded policy iteration (BPI) algorithm (see [14] for details) and
describe a simple extension to bias its search toward reachable belief states. BPI incrementally constructs an FSC by alternating policy improvement and policy evaluation. Unlike policy iteration [7], this is done by slowly increasing the number of nodes (and value
vectors). The policy improvement step greedily improves each node n by optimizing its
action and observation strategies by solving the linear program (LP) in Table 1. This LP
uniformly maximizes the improvement in the value function by optimizing n?s distributions Pr(a, n0 |n, z). The policy evaluation step computes the value function of the current
controller by solving Eq. 1. The algorithm monotonically improves the policy until convergence to a local optimum, at which point new nodes are introduced to escape the local
optimum. BPI is guaranteed to converge to a policy that is optimal at the ?tangent? belief
states while slowly growing the size of the controller [14].
In practice, we often wish to find a policy suitable for a given initial belief state. Since
only a small subset of belief space is often reachable, it is generally possible to construct
much smaller policies tailored to the reachable region. We now describe a simple way to
bias BPI?s efforts toward the reachable region. Recall that the LP in Table 1 optimizes the
parameters of a node to uniformly improve its value at all belief states. We propose a new
LP (Table 2) that weighs the improvement by the (unnormalized) discounted occupancy
distribution induced by the current policy. This accounts for belief states reachable for the
node by aggregating them together. The (unnormalized) discounted occupancy distribution
is given by:
o(s0 , n0 ) = b0 (s0 , n0 ) + ?
X
o(s, n) Pr(a|n) Pr(z|a, s) Pr(n0 |n, z)
?s0 , n0
s,a,z,n
The LP in Table
P 2 is obtained by introducing variables s,n for each s, replacing the objective by s,n o(s, n)s,n and replacing in each constraint by the corresponding s,n .
When using the modified LP, BPI naturally tries to improve the policy at the reachable belief states before the others. Since the modification ensures that the value function doesn?t
decrease at any belief state, focusing the efforts on reachable belief states won?t decrease
policy value at other belief states. Furthermore, though the policy is initially biased toward
reachable states, BPI will eventually improve the policy for all belief states.
f
b
T?
~?
T
~
b
b?
~?
T
~
b?
~
R
R
T?
~
R
R
r
r?
Figure 1: Functional flow of a POMDP (dotted arrows) and a compressed POMDP (solid
arrows).
4 Value-Directed Compression
We briefly review the sufficient conditions for a lossless compression of POMDPs [13] and
describe a simple new algorithm to obtain good lossy compressions. Belief states constitute
a sufficient statistic summarizing all information available to the decision maker (i.e., past
actions and observations). However, as long as enough information is available to evaluate
the value of each policy, one can still choose the best policy. Since belief states often
contain information irrelevant to the estimation of future rewards, one can often compress
belief states into some lower-dimensional representation. Let f be a compression function
that maps each belief state b into some lower dimensional compressed belief state ?b (see
Figure 1). Here ?b can be viewed as a bottleneck that filters the information contained in
b before it is used to estimate future rewards. We desire a compression f such that ?b
corresponds to the smallest statistic sufficient for accurately predicting the current reward
r as well as the next compressed belief state ?b0 (since it captures all the information in b0
necessary to accurately predict subsequent rewards). Such a compression f exists if we can
? such
also find compressed transition dynamics T? a,z and a compressed reward function R
that:
? ? f and f ? T a,z = T? a,z ? f ?a ? A, z ? Z
R=R
(3)
a,z
?
?
Given an f , R and T satisfying Eq. 3, we can evaluate any policy ? using the compressed
POMDP dynamics to obtain V? ? . Since V ? = V? ? ?f , the compressed POMDP is equivalent
to the original.
When restricting f to be linear (represented by matrix F ), we can rewrite Eq. 3
? and T a,z F = F T? a,z ?a ? A, z ? Z
R = FR
(4)
a,z
That is, the column space of F spans R and is invariant w.r.t. each T . Hence, the
columns of the best linear lossless compression mapping F form a basis for the smallest
invariant subspace (w.r.t. each T a,z ) that spans R, i.e., the Krylov subspace. We can find the
columns of F by Krylov iteration: multiplying R by each T a,z until the newly generated
vectors are linear combinations of previous ones.1 The dimensionality of the compressed
space is equal to the number of columns of F , which is necessarily smaller than or equal
? and
to the dimensionality of the original belief space. Once F is found, we can compute R
a,z
?
each T
by solving the system in Eq. 4.
Since linear lossless compressions are not always possible, we can extend the technique of
[13] to find good lossy compressions with early stopping of the Krylov iteration. We retain
only the vectors that are ?far? from being linear combinations of prior vectors. For instance,
if v is a linearP
combination of v1 , v2 , . . . , vn , then there are coefficients c1 , c2 , . . . , cn s.t.
the error ||v ? i ci vi ||2 is zero. Given a threshold or some upper bound k on the desired
number of columns in F , we run Krylov iteration, retaining only the vectors with an error
greater than , or the k vectors with largest error. When F is computed by approximate
1
For numerical stability, one must orthogonalize each vector before multiplying by T a,z .
? and T? a,z by solving the linear system in Eq. 4?
Krylov iteration, we cannot compute R
due to the lossy nature of the compression, the system is overconstrained. But we can find
? and T? a,z by computing a least square approximation, solving:
suitable R
? and F > T a,z F = F > F T? a,z ?a ? A, z ? Z
F >R = F >F R
While compression is required when the dimensionality of belief space is too large, unfortunately, the columns of F have the same dimensionality. Factored POMDPs of exponential
dimension can, however, admit practical Krylov iteration if carried out using a compact
? and each T? a,z .
representation (e.g., DTs or ADDs) to efficiently compute F , R
5 Bounded Policy Iteration with Value-Directed Compression
In principle, any POMDP algorithm can be used to solve the compressed POMDPs produced by VDC. If the compression is lossless and the POMDP algorithm exact, the computed policy will be optimal for the original POMDP. In practice, POMDP algorithms are
usually approximate and lossless compressions are not always possible, so care must be
taken to ensure numerical stability and a policy of high quality for the original POMDP.
We now discuss some of the integration issues that arise when combining VDC with BPI.
Since V = F V? , maximizing the compressed value vector V? of some node n automatically
maximizes the value V of n w.r.t. the original POMDP when F is nonnegative; hence it is
essential that F be nonnegative. Otherwise, the optimal policy of the compressed POMDP
may not be optimal for the original POMDP. Fortunately, when R is nonnegative then F
is guaranteed to be nonnegative by the nature of Krylov iteration. If some rewards are
negative, we can add a sufficiently large constant to R to make it nonnegative without
changing the decision problem.
Since most algorithms, including BPI, compute approximately optimal policies it is also
critical to normalize the columns of F . Suppose F has two columns f 1 and f2 with L1 lengths 1 and 100, respectively. Since V = F V? = v?1 f1 + v?2 f2 , changes in v?2 have a
much greater impact on V than changes in v?1 . Such a difference in sensitivity may bias the
search for a good policy to an undesirable region of the belief space, or may even cause
the algorithm to return a policy that is far from optimal for the original POMDP despite the
fact that it is -optimal for the compressed POMDP.
We note that it is ?safer? to evaluate policies iteratively by successive approximation rather
than solving the system in Eq. 1. By definition, the transition matrices T a,z have eigenvalues with magnitude ? 1. In contrast, lossy compressed transition matrices T? a,z are not
guaranteed to have this property. Hence, solving the system in Eq. 1 may not correspond
to policy evaluation. It is thus safer to evaluate policies by successive approximation for
lossy compressions.
Finally several algorithms including BPI compute witness belief states to verify the dominance of a value vector. Since the compressed belief space B? is different from the original
belief space B, this must be approached with care. B is a simplex corresponding to the
convex hull of the state points. In contrast, since each row vector of F is the compressed
version of some state point, B? corresponds to the convex hull of the row vectors of F .
When F is non-negative, it is often possible to ignore this difference. For instance, when
verifying the dominance of a value vector, if there is a compressed witness ?b, there is always an uncompressed witness b, but not vice-versa. This means that we can properly
identify all dominating value vectors, but we may erroneously classify a dominated vector
as dominating. In practice, this doesn?t impact the correctness of algorithms such as policy
iteration, bounded policy iteration, incremental pruning, witness algorithm, etc. but it will
slow them down since they won?t be able to prune as many value vectors as possible.
cycle16
cycle19
100
95
90
250
Expected Rewards
120
Expected Rewards
Expected Rewards
105
cycle22
115
110
105
250
200
150
100
50
# of basis fns
20
40
60
100
80
150
100
50
# of basis fns
# of nodes
cycle25
20
40
60
80
100
200
120
150
100
140
135
110
105
100
150
50
40
60
100
80
150
100
50
# of basis fns
# of nodes
100
120
# of nodes
130
125
120
115
20
40
60
80
100
200
120
150
100
50
# of basis fns
# of nodes
3legs25
3legs22
20
40
60
80
100
120
# of nodes
cycle25
150
145
140
135
130
Time (1000 seconds)
12
Expected Rewards
Expected Rewards
80
250
200
120
60
110
250
200
40
135
115
95
250
20
3legs19
Expected Rewards
Expected Rewards
Expected Rewards
145
# of basis fns
50
# of basis fns
# of nodes
120
20
120
3legs16
150
100
125
250
200
120
130
160
155
150
145
140
250
125
250
200
150
100
# of basis fns
50
20
40
60
80
100
# of nodes
120
10
8
6
4
2
250
200
150
100
# of basis fns
50
20
40
60
80
100
120
200
150
100
# of basis fns
# of nodes
50
20
40
60
80
100
120
# of nodes
Figure 2: Experimental results for cycle and 3legs network configurations of 16, 19,
22 and 25 machines. The bottom right graph shows the running time of BPI on compressed
versions of a cycle network of 25 machines.
VDCBPI
heuristic
doNothing
16
120.9
100.6
98.4
3legs
19
22
137.0 151.0
118.3 138.3
112.9 133.5
25
164.8
152.3
147.1
16
103.9
102.5
91.6
cycle
19
22
121.3 134.3
117.9 130.2
105.4 122.0
25
151.4
152.3
140.1
Table 3: Comparison of the best policies achieved by VDCBPI to the doNothing and
heuristic policies.
The above tips work well when VDC is integrated with BPI. We believe they are sufficient
to ensure proper integration of VDC with other POMDP algorithms, though we haven?t
verified this empirically.
6 Experiments
We report on experiments with VDCBPI on some synthetic network management problems
similar to those introduced in [5]. A system administrator (SA) maintains a network of
machines. Each machine has a 0.1 probability of failing at any stage; but this increases to
0.333 when a neighboring machine is down. The SA receives a reward of 1 per working
machine and 2 per working server. At each stage, she can either reboot a machine, ping a
machine or do nothing. She only observes the status of a machine (with 0.95 accuracy) if
she reboots or pings it. Costs are 2.5 (rebooting), 0.1 (pinging), and 0 (doing nothing). An
n-machine network induces to a POMDP with 2n states, 2n + 1 actions and 2 observations.
We experimented with networks of 16, 19, 22 and 25 machines organized in two configurations: cycle (a ring) and 3legs (a tree of 3 branches joined at the root). Figure 2 shows
the average expected reward earned by policies computed by BPI after the POMDP has
been compressed by VDC. Results are averaged over 500 runs of 60 steps, starting with a
belief state where all machines are working.2 As expected, decision quality increases as we
increase the number of nodes used in BPI and basis functions used in VDC. Also interesting
are some of the jumps in the reward surface of some graphs, suggesting phase transitions
w.r.t. the dimensionality of the compression. The bottom right graph in Fig. 2 shows the
time taken by BPI on a cycle network of 25 machines (other problems exhibit similar
behavior). VDC takes from 4902s. to 12408s. (depending on size and configuration) to
compress POMDPs to 250 dimensions.3
In Table 3 we compare the value of the best policy with less than 120 nodes found by
VDCBPI to two other simple policies. The doNothing policy lets the network evolve
without any rebooting or pinging. The heuristic policy estimates at each stage the
probability of failure4 of each machine and reboots the machine most likely to be down
if its failure probability is greater than threshold p1 or pings it if greater than threshold
p2 . Settings of p1 = 0.8 and p2 = 0.15 were used.5 This heuristic policy performs very
well and therefore offers a strong competitor to VDCBPI. But it is possible to do better
than the heuristic policy by optimizing the choice of the machine that the SA may reboot
or ping. Since a machine is more likely to fail when neighboring machines are down, it
is sometimes better to choose (for reboot) a machine surrounded by working machines.
However, since the SA doesn?t exactly know which machines are up or down due to partial
observability, such a tradeoff is difficult to evaluate and sometimes not worthwhile. With a
sufficient number of nodes and basis functions, VDCBPI outperforms the heuristic policy
on the 3legs networks and matches it on the cycle networks. This is quite remarkable
given the fact that belief states were compressed to 250 dimensions or less compared to the
original dimensionality ranging from 65,536 to 33,554,432.
7 Conclusion
We have described a new POMDP algorithm that mitigates both high belief space dimensionality and policy/VF complexity. By integrating value-directed compression with
2
The ruggedness of the graphs is mainly due to the variance in the reward samples.
Reported running times are the cputime measured on 3GHz linux machines.
4
Due to the large state space, approximate monitoring was performed by factoring the joint.
5
These values were determined through enumeration of all threshold combinations in increments
of 0.05, choosing the best for 25-machine problems.
3
bounded policy iteration, we can solve synthetic network management POMDPs of 33 million states (3 orders of magnitude larger than previously solved discrete POMDPs). Note
that the scalability of VDCBPI is problem dependent, however we hope that new, scalable, approximate POMDP algorithms such as VDCBPI will allow POMDPs to be used to
model real-world problems, with the expectation that they can be solved effectively. We
also described several improvements to the existing VDC and BPI algorithms.
Although VDC offers the advantage that any existing solution algorithm can be used to
solve compressed POMDPs, it would be interesting to combine BPI or PBVI with a factored representation such as DTs or ADDs, allowing one to directly solve large scale
POMDPs without recourse to an initial compression. Beyond policy space complexity
and high dimensional belief spaces, further research will be necessary to deal with exponentially large action and observation spaces.
References
[1] D. Aberdeen and J. Baxter. Scaling internal-state policy-gradient methods for POMDPs. Proc.
of the Nineteenth Intl. Conf. on Machine Learning, pp.3?10, Sydney, Australia, 2002.
[2] C. Boutilier and D. Poole. Computing optimal policies for partially observable decision processes using compact representations. Proc. AAAI-96, pp.1168?1175, Portland, OR, 1996.
[3] D. Braziunas and C. Boutilier. Stochastic local search for POMDP controllers. Proc. AAAI-04,
to appear, San Jose, CA, 2004.
[4] A. R. Cassandra, M. L. Littman, and N. L. Zhang. Incremental pruning: A simple, fast, exact
method for POMDPs. Proc. UAI-97, pp.54?61, Providence, RI, 1997.
[5] C. Guestrin, D. Koller, and R. Parr. Max-norm projections for factored MDPs. Proc. IJCAI-01,
pp.673?680, Seattle, WA, 2001.
[6] C. Guestrin, D. Koller, and R. Parr. Solving factored POMDPs with linear value functions.
IJCAI-01 Wkshp. on Planning under Uncertainty and Incomplete Information, Seattle, 2001.
[7] E. A. Hansen. Solving POMDPs by searching in policy space. Proc. UAI-98, pp.211?219,
Madison, Wisconsin, 1998.
[8] E. A. Hansen and Z. Feng. Dynamic programming for POMDPs using a factored state representation. Proc. AIPS-2000, pp.130?139, Breckenridge, CO, 2000.
[9] E. A. Hansen and Z. Feng. Approximate planning for factored POMDPs. Proc. ECP-2001,
Toledo, Spain, 2000.
[10] L. P. Kaelbling, M. Littman, and A. R. Cassandra. Planning and acting in partially observable
stochastic domains. Artif. Intel., 101:99?134, 1998.
[11] N. Meuleau, L. Peshkin, K. Kim, and L. P. Kaelbling. Learning finite-state controllers for
partially observable environments. Proc. UAI-99, pp.427?436, Stockholm, 1999.
[12] J. Pineau, G. Gordon, and S. Thrun. Point-based value iteration: an anytime algorithm for
POMDPs. IJCAI-03, Acapulco, Mexico, 2003.
[13] P. Poupart and C. Boutilier. Value-directed compressions of POMDPs. Advances in Neural
Information Processing Systems, pp.1547?1554, Vancouver, Canada, 2002.
[14] P. Poupart and C. Boutilier. Bounded finite state controllers. Advances in Neural Information
Processing Systems, Vancouver, Canada, 2003.
[15] N. Roy and G. Gordon. Exponential family PCA for belief compression in pomdps. Advances
in Neural Information Processing Systems, pp.1635?1642, Vancouver, BC, 2002.
[16] M. T. J. Spaan and N. Vlassis. A point-based pomdp algorithm for robot planning. IEEE Intl.
Conf. on Robotics and Automation, to appear, New Orleans, 2004.
| 2704 |@word version:2 briefly:2 compression:24 norm:1 solid:1 carry:1 initial:3 cyclic:1 configuration:3 bc:1 past:1 existing:4 outperforms:1 current:4 si:1 must:3 numerical:2 subsequent:1 update:1 n0:15 meuleau:1 node:26 toronto:6 successive:2 zhang:1 c2:1 combine:2 expected:10 behavior:1 p1:2 nor:1 growing:1 planning:4 roughly:1 bellman:1 discounted:3 automatically:1 curse:2 enumeration:1 increasing:1 spain:1 bounded:9 maximizes:2 didn:1 reboot:3 mitigate:1 rebooting:2 exactly:1 appear:2 before:3 local:5 aggregating:1 despite:1 approximately:1 conversely:1 co:1 limited:1 averaged:1 directed:9 practical:1 orleans:1 practice:4 projection:1 integrating:1 cannot:1 ga:2 undesirable:1 equivalent:1 map:1 demonstrated:1 maximizing:1 starting:1 convex:3 pomdp:27 factored:7 rule:1 classic:2 stability:2 searching:1 increment:1 updated:1 suppose:1 exact:2 programming:1 associate:2 roy:1 satisfying:1 labeled:1 observed:1 bottom:2 solved:3 capture:1 verifying:1 thousand:1 region:3 compressing:1 ensures:1 cycle:6 earned:1 decrease:2 observes:1 environment:1 complexity:8 reward:20 littman:2 dynamic:3 solving:9 rewrite:1 efficiency:1 f2:2 basis:13 joint:1 represented:3 various:1 fast:1 describe:4 fsc:3 approached:1 choosing:1 whose:1 heuristic:7 larger:3 solve:7 dominating:2 nineteenth:1 quite:1 otherwise:1 compressed:22 ability:2 statistic:2 itself:1 advantage:2 eigenvalue:1 propose:2 fr:1 neighboring:2 combining:3 pbvi:3 ppoupart:1 normalize:1 scalability:4 seattle:2 convergence:1 ijcai:3 optimum:3 p:1 intl:2 produce:1 incremental:2 executing:2 ring:1 depending:1 measured:1 b0:3 sa:4 strong:1 sydney:1 implemented:1 c:2 eq:7 p2:2 filter:1 stochastic:5 hull:2 australia:1 successor:1 f1:1 acapulco:1 stockholm:1 extension:1 sufficiently:1 mapping:2 predict:1 parr:2 early:1 smallest:2 failing:1 estimation:1 proc:9 hansen:4 maker:1 largest:1 vice:1 correctness:1 hope:1 always:3 modified:1 rather:1 avoid:1 focus:1 improvement:5 properly:1 she:3 portland:1 mainly:1 braziunas:1 contrast:2 greedily:1 kim:1 summarizing:1 abstraction:1 stopping:1 factoring:1 dependent:1 typically:1 integrated:1 initially:1 koller:2 mitigating:1 issue:1 among:2 pascal:1 augment:1 retaining:1 integration:2 equal:2 once:2 construct:2 uncompressed:1 future:2 simplex:1 others:1 report:1 piecewise:1 escape:1 few:2 distinguishes:1 haven:1 gordon:2 simultaneously:1 phase:1 evaluation:3 devoted:1 edge:2 partial:1 necessary:3 tree:2 incomplete:1 desired:1 weighs:1 instance:2 column:8 classify:1 cost:1 introducing:1 kaelbling:2 subset:1 uniform:1 too:1 reported:1 providence:1 synthetic:4 sensitivity:1 retain:1 tip:1 together:1 linux:1 aaai:2 management:4 hn:1 possibly:1 slowly:2 choose:2 administrator:1 bpi:23 admit:1 conf:2 return:1 account:1 suggesting:1 automation:1 coefficient:1 vi:1 performed:1 try:1 root:1 observing:1 doing:1 bayes:1 maintains:1 contribution:1 square:1 accuracy:1 variance:1 efficiently:1 correspond:2 identify:1 accurately:2 craig:1 produced:1 none:1 monitoring:2 pomdps:26 multiplying:2 m5s:2 ping:4 definition:1 competitor:1 failure:1 pp:9 naturally:1 associated:1 newly:1 recall:1 knowledge:1 anytime:1 dimensionality:13 improves:2 organized:1 overconstrained:1 focusing:1 response:1 done:1 though:2 furthermore:1 stage:3 until:2 working:4 receives:1 replacing:2 ei:1 expressive:1 lack:1 incrementally:1 pineau:1 quality:3 believe:1 lossy:6 artif:1 contain:1 verify:1 hence:5 alternating:1 iteratively:1 deal:2 steady:1 unnormalized:3 won:2 demonstrate:2 performs:1 l1:1 ranging:1 functional:1 empirically:1 exponentially:3 million:3 extend:1 slight:1 versa:1 breckenridge:1 reachable:9 fscs:1 robot:1 surface:1 etc:1 add:5 optimizing:3 optimizes:1 irrelevant:1 server:1 guestrin:2 greater:4 care:2 fortunately:1 prune:1 converge:1 monotonically:1 branch:1 match:1 offer:4 long:1 wkshp:1 impact:2 scalable:4 controller:7 expectation:1 bz:1 iteration:18 sometimes:2 tailored:1 achieved:1 robotics:1 c1:1 background:1 addressed:1 diagram:1 source:5 biased:1 unlike:1 ascent:1 induced:1 flow:1 near:1 enough:1 baxter:1 zi:1 reduce:1 observability:1 cn:1 tradeoff:1 bottleneck:1 peshkin:1 pca:3 effort:2 algebraic:1 cause:1 constitute:1 action:10 boutilier:5 generally:1 discount:1 induces:1 sl:2 dotted:1 per:2 discrete:4 dominance:2 threshold:4 changing:1 verified:1 v1:1 graph:5 run:2 jose:1 uncertainty:1 h5:2 family:1 vn:1 ecp:1 decision:9 scaling:1 vf:4 capturing:1 bound:1 guaranteed:3 tackled:1 nonnegative:5 constraint:1 ri:1 dominated:1 erroneously:1 span:2 department:2 according:1 combination:6 representable:1 describes:1 slightly:1 smaller:3 lp:8 spaan:1 making:1 modification:1 leg:4 invariant:2 pr:33 fns:9 taken:2 recourse:1 equation:1 previously:1 discus:1 eventually:1 fail:1 know:1 fed:1 available:2 worthwhile:1 v2:1 indirectly:1 original:9 compress:2 denotes:1 running:2 cf:3 ensure:2 madison:1 exploit:1 feng:3 objective:1 realized:1 strategy:3 exhibit:1 gradient:2 subspace:3 thrun:1 poupart:3 toward:3 length:1 mexico:1 difficult:1 unfortunately:1 negative:2 rise:1 proper:1 policy:62 allowing:2 upper:1 observation:11 markov:2 finite:3 immediate:1 witness:4 vlassis:1 canada:2 introduced:2 pair:1 required:2 plague:1 cputime:1 dts:4 toledo:1 address:1 beyond:2 able:1 krylov:7 usually:1 poole:1 program:1 max:4 including:2 belief:40 suitable:2 critical:1 natural:1 predicting:1 occupancy:3 improve:5 lossless:5 mdps:1 carried:1 review:2 prior:1 tangent:1 evolve:1 vancouver:3 wisconsin:1 interesting:2 ruggedness:1 remarkable:1 agent:1 sufficient:5 s0:11 principle:1 intractability:5 surrounded:1 row:2 cebly:1 bias:4 allow:1 ghz:1 dimension:3 transition:6 world:2 computes:1 doesn:3 jump:1 san:1 far:2 sj:2 approximate:10 observable:5 pruning:5 compact:4 ignore:1 status:1 uai:3 search:5 vdc:14 table:8 nature:2 ca:1 necessarily:1 domain:1 arrow:2 arise:1 nothing:2 fig:1 intel:1 slow:1 wish:1 exponential:8 down:5 specific:1 mitigates:2 experimented:1 exists:1 essential:1 restricting:1 effectively:1 ci:1 magnitude:2 horizon:1 cassandra:2 aberdeen:1 likely:2 desire:1 adjustment:1 contained:1 partially:5 joined:1 corresponds:2 satisfies:1 viewed:2 consequently:1 change:2 safer:2 infinite:2 determined:1 uniformly:3 acting:1 experimental:1 orthogonalize:1 internal:1 evaluate:5 |
1,878 | 2,705 | A Generalized Bradley-Terry Model: From
Group Competition to Individual Skill
Tzu-Kuo Huang
Chih-Jen Lin
Department of Computer Science
National Taiwan University
Taipei 106, Taiwan
Ruby C. Weng
Department of Statistics
National Chenechi University
Taipei 116, Taiwan
Abstract
The Bradley-Terry model for paired comparison has been popular in
many areas. We propose a generalized version in which paired individual
comparisons are extended to paired team comparisons. We introduce a
simple algorithm with convergence proofs to solve the model and obtain
individual skill. A useful application to multi-class probability estimates
using error-correcting codes is demonstrated.
1
Introduction
The Bradley-Terry model [2] for paired comparisons has been broadly applied in many
areas such as statistics, sports, and machine learning. It considers the model
?i
P (individual i beats individual j) =
,
(1)
?i + ? j
where ?i is the overall skill of the ith individual. Given k individuals and rij as the number
of times that i beats j, an approximate skill pi can be found by minimizing the negative log
likelihood of the model (1):
?
X?
pi
pj
min
l(p) = ?
rij log
+ rji log
p
pi + p j
pi + p j
i<j
subject to
0 ? pi , i = 1, . . . , k,
k
X
pi = 1.
(2)
i=1
Thus, from paired comparisons, we can obtain individual performance. This model dates
back to [14] and has been extended to more general settings. Some reviews are, for example, [5, 6]. Problem (2) can be solved by a simple iterative procedure:
Algorithm 1
1. Start with any initial p0j > 0, j = 1, . . . , k.
2. Repeat (t = 0, 1, . . .)
a. Let s = (t mod k) + 1. For j = 1, . . . , k, define
? P
si
? P i:i6=rs r+r
if j = s,
si
is
t,n
i:i6=s pt +pt
pj ?
s
i
? t
pj
if j 6= s.
(3)
b. Normalize pt,n to be pt+1 .
until ?l(pt )/?pj = 0, j = 1, . . . , k are satisfied.
This algorithm is so simple that there is no need to use sophisticated optimization techniques. If rij > 0, ?i, j, Algorithm 1 globally converges to the unique minimum of (2). A
systematic study of the convergence is in [9].
Several machine learning work have used the Bradley-Terry model and one is to obtain
multi-class probability estimates from pairwise coupling [8]. For any data instance x, if
nij is the number of training data in the ith or jth class, and
rij ? nij P (x in class i | x in class i or j)
is available, solving (2) obtains the estimate of P (x in class i), i = 1, . . . , k. [13] tried to
extend this algorithm to other multi-class settings such as ?one-against-the rest? or ?errorcorrecting coding,? but did not provide a convergence proof. In Section 5.2 we show that
the algorithm proposed in [13] indeed has some convergence problems.
In this paper, we propose a generalized Bradley-Terry model where each comparison is
between two disjoint subsets of subjects. Then from the results of team competitions, we
can approximate the skill of each individual. This model has many potential applications.
For example, from records of tennis or badminton doubles (or singles and doubles combined), we may obtain the rank of all individuals. A useful application in machine learning
is multi-class probability estimates using error-correcting codes. We then introduce a simple iterative method to solve the generalized model with a convergence proof. Experiments
on multi-class probability estimates demonstrate the viability of the proposed model and
algorithm. Due to space limitation, we omit all proofs in this paper.
2
Generalized Bradley-Terry Model
We propose a generalized Bradley-Terry model where, using team competition results, we
can approximate individual skill levels. Consider a group of k individuals: {1, . . . , k}.
Two disjoint subsets Ii+ and Ii? form teams for games and ri ? 0 (ri? ? 0) is the number
of times that Ii+ beats Ii? (Ii? beats Ii+ ). Thus, we have Ii ? {1, . . . , k}, i = 1, . . . , m so
that
Ii = Ii+ ? Ii? ,
Ii+ 6= ?, Ii? 6= ?, and Ii+ ? Ii? = ?.
Under the model that
P
P
j?Ii+ ?j
j?I + ?j
+
?
P
P (Ii beats Ii ) = P
= P i
,
j?Ii ?j
j?I + ?j +
j?I ? ?j
i
we can define
qi ?
X
pj ,
j?Ii
qi+ ?
i
X
j?Ii+
pj ,
qi? ?
X
pj ,
j?Ii?
and minimize the negative log likelihood
min
p
l(p) = ?
m
X
?
?
ri log(qi+ /qi ) + ri? log(qi? /qi ) ,
i=1
under the same constraints of (2). If Ii , i = 1, . . . , k(k ? 1)/2 are as the following:
Ii+
Ii?
ri
ri?
{1}
{2} r12
r21
..
..
..
..
.
.
.
.
{k ? 1} {k} rk?1,k rk,k?1
(4)
then (4) goes back to (2). The difficulty of solving (4) over solving (2) is that now l(p)
is expressed in terms of qi+ , qi? , qi but the real variable is p. The original Bradley-Terry
model is a special case of other statistical models such as log-linear or generalized linear
model, so methods other than Algorithm 1 (e.g., iterative scaling and iterative weighted
least squares) can also be used. However, (4) is not in a form of such models and hence
these methods cannot be applied. We propose the following algorithm to solve (4).
Algorithm 2
1. Start with p0j > 0, j = 1, . . . , k and corresponding qi0,+ , qi0,? , qi0 , i = 1, . . . , m.
2. Repeat (t = 0, 1, . . .)
a. Let s = (t mod k) + 1. For j = 1, . . . , k, define
?P
P
ri?
ri
?
? i:s?Ii+ qit,+ + i:s?Ii? qit,? t
pj if j = s,
P
ri +r ?
i
(5)
pt,n
j ?
i:s?Ii
qt
?
i
? t
pj
if j 6= s.
b. Normalize pt,n to pt+1 .
c. Update qit,+ , qit,? , qit to qit+1,+ , qit+1,? , qit+1 , i = 1, . . . , m.
until ?l(pt )/?pj = 0, j = 1, . . . , k are satisfied.
For the multiplicative factor in (5) to be well defined (i.e., non-zero denominator), we
need Assumption 1, which will be discussed in Section 3. Eq. (5) is a simple fixed-point
type update; in each iteration, only one component (i.e., pts ) is modified while the others
remain the same. It is motivated from using a descent direction to strictly decrease l(p): If
?l(pt )/?ps 6= 0, then
? ?
!
?2 ! . ? X
?l(pt )
?l(pt )
ri + ri?
t,n
t
t
< 0,
(6)
? (ps ? ps ) = ?
ps
?ps
?ps
qit
i:s?Ii
where
?l(p)
?ps
pt,n
s
= ?
X r?
X ri + r?
X ri
i
i
.
+ ?
? +
q
q
q
i
i
i
?
+
i:s?I
i:s?Ii
i:s?Ii
i
pts
? is a descent direction in optimization since a sufficiently small step along
Thus,
this direction guarantees the strict decrease of the function value. Since now we take the
whole direction without searching for the step size, more efforts are needed to prove the
strict decrease in Lemma 1. However, (6) does hint that (5) is a reasonable update.
Lemma 1 If pts > 0 is the index to be updated and ?l(pt )/?ps 6= 0, then l(pt+1 ) < l(pt ).
If we apply the update rule (5) on the pairwise model,
P
P
rsi
i:i6=s pts
i:i6=s rsi
t
P
P
rsi
ris ps = P
rsi +ris and (5) goes back to (3).
i:i6=s pt +pt +
i:i6=s pt +pt
i:i6=s pt +pt
s
3
i
s
i
s
i
Convergence of Algorithm 2
For any point satisfying ?l(p)/?pj = 0, j = 1, . . . , k and constraints of (4), it is a stationary point of (4)1 . We will prove that Algorithm 2 converges to such a point. If
1
A stationary point means a Karash-Kunh-Tucker (KKT) point for constrained optimization problems like (2) and (4). Note that here ?l(p)/?pj = 0 implies (and is more restricted than) the KKT
condition.
it stops in a finite number of iterations, then ?l(p)/?pj = 0, j = 1, . . . , k, which
means a stationary point of (4) is already obtained. Thus, we only need to handle the
case where {pt } is an infinite sequence. As {pt }?
t=0 is in a compact (i.e., closed and
Pk
bounded) set {p | 0 ? pj ? 1, j=1 pj = 1}, it has at least one convergent subsequence. Assume p? is one such convergent point. In the following we will prove that
?l(p? )/?pj = 0, j = 1, . . . , k.
To prove the convergence of a fixed-point type algorithm, we need that if p?s > 0 and
?l(p? )/?ps 6= 0, then from p?s we can use (5) to update it to p?,n
6= p?s . We thus make the
s
following assumption to ensure that p?s > 0 (see also Theorem 1).
Assumption 1 For each j ? {1, . . . , k},
?i:i?A Ii = {1, . . . , k}, where A = {i | (Ii+ = {j}, ri > 0) or (Ii? = {j}, ri? > 0)}.
That is, each individual forms a winning (losing) team in some competitions which together
involve all subjects.
An issue left in Section 2 is whether the multiplicative factor in (5) is well defined. With
Assumption 1 and initial p0j > 0, j = 1, . . . , k, one can show by induction that ptj > 0, ?t
and hence the denominator of (5) is never zero: If ptj > 0, Assumption 1 implies that
P
P
t,+
or i:j?I ? ri? /qit,? is positive. Thus, both numerator and denominator in
i:j?I + ri /qi
i
i
the multiplicative factor are positive, and so is pt+1
j .
If rij > 0, the original Bradley-Terry model satisfies Assumption 1. No matter the model
satisfies the assumption or not, an easy way to fulfill it is to add an additional term
!
?
k
X
ps
(7)
??
log Pk
s=1
j=1 pj
to l(p), where ? is a small positive number. That is, for each s, we make an Ii = {1, . . . , k}
Pk
with Ii+ = {s}, ri = ?, and ri? = 0. As j=1 pj = 1 is one of the constraints, (7) reduces
Pk
to ?? s=1 log ps , which is a barrier term in optimization to ensure that ps does not go to
zero. The property p?s > 0 and the convergence of Algorithm 2 are in Theorem 1:
Theorem 1 Under Assumption 1, any convergent point p? of Algorithm 2 satisfies p?s >
0, s = 1, . . . , k and is a stationary point of (4).
4
Asymptotic Distribution of the Maximum Likelihood Estimator
For the standard Bradley-Terry model, asymptotic distribution of the MLE (i.e., p) has been
discussed in [5]. In this section, we discuss the asymptotic distribution for the proposed
estimator. To work on
?, we define
P the real probability
P
P
q?i ? j?Ii ?j ,
q?i+ ? j?I + ?j ,
q?i? ? j?I ? ?j ,
i
i
and consider ni ? ri + ri? as a constant. Note that ri ? BIN(ni , q?i+ /?
qi ) is a random
variable representing the number of times that Ii+ beats Ii? in ni competitions. By defining
for s, t = 1, . . . , k,h
i P
P
n q??
n q?+
= i:s?I + q?+i q?i2 + i:s?I ? q??i q?i2 ,
?ss ? var ?l(?)
?ps
i
i
i i
h
i Pi i
q?? n
?l(?)
?st ? cov ?l(?)
= i:s,t?I + q?i+ q?2i ?
?ps , ?pt
i
i i
P
P
P
q?i+ ni
ni
ni
+
?
?
i:(s,t)?Ii ?Ii q?i2
i:(s,t)?Ii? ?Ii+ q?i2 +
i:s,t?Ii? q?? q?2 , s 6= t,
i i
we have the following theorem:
Theorem
? 2 Let n be the?total number of comparisons. If ri is independent of rj , ?i 6= j,
then n(p1 ? ?1 ), . . . , n(pk?1 ? ?k?1 ) have for large samples the multivariate normal
distribution with zero means and dispersion matrix [??st ]?1 , where
??st = ?st ? ?sk ? ?tk + ?kk , s, t = 1, . . . , k ? 1.
5
Application to Multi-class Probability Estimates
Many classification methods are two-class based approaches and there are different ways
to extend them for multi-class cases. Most existing studies focus on predicting class labels
but not probability estimates. In this section, we discuss how the generalized Bradley-Terry
model can be applied to multi-class probability estimates.
Error-correction coding [7, 1] is a general method to construct binary classifiers and combine them for multi-class prediction. It suggests some ways to construct Ii+ and Ii? ; both
are subsets of {1, . . . , k}. Then one trains a binary model using data from classes in Ii+
(Ii? ) as positive (negative). Simple and commonly used methods such as ?one-against-one?
and ?one-against-the rest? are its special cases. Given ni the number of training data with
classes in Ii = Ii+ ? Ii? , we assume here that for any data x,
ri ? ni P (x in classes of Ii+ | x in classes of Ii+ or Ii? )
(8)
is available, and the task is to approximate P (x in class s), s = 1, . . . , k. In the rest of this
section we discuss the special case ?one-against-the rest? and the earlier results in [13].
5.1
Properties of the ?One-against-the rest? Approach
For this approach, Ii , i = 1, . . . , m are
Ii+
Ii?
{1} {2, . . . , k}
{2} {1, 3, . . . , k}
..
..
.
.
ri
r1
r2
..
.
ri?
1 ? r1
1 ? r2
..
.
Now n1 = ? ? ? = nm = the total number of training data, so the solution to (4) is not
affected by ni . Thus, we remove it from (8), so ri + ri? = 1. As ?l(p)/?ps = 0 becomes
k
X
rj?
r1 1 ? r1
rk 1 ? rk
rs X rj?
+
= k, we have ?
= ??? =
?
= k?
= ?,
ps
1 ? pj
p1 1 ? p 1
pk 1 ? p k
1 ? pj
j=1
j:j6=s
wherep? is a constant. These equalities provide another
p way to solve p, and ps = ((1 +
?) ? (1 + ?)2 ? 4rs ?)/2?. Note that ((1 + ?) + (1 + ?)2 ? 4rs ?)/2? also satisfies
the equalities, but it is negative when ? < 0, and greater than 1 when ? > 0. By solving
P
m
s=1 ps = 1, we obtain ? and the optimal p.
From the formula of ps , if ? > 0, larger ps implies smaller (1 + ?)2 ? 4rs ? and hence larger
rs . It is similar for ? < 0. Thus, the order of p1 , . . . , pk is the same as that of r1 , . . . , rk :
Theorem 3 If rs ? rt , then ps ? pt .
5.2
The Approach in [13] for Error-Correcting Codes
[13] was the first attempt to address the probability estimates using general error-correcting
codes. By considering the same optimization problem (4), it proposes a heuristic update
rule
P
P
?
i:s?Ii+ ri +
i:s?Ii? ri
t,n
ps ? P
pt ,
(9)
P
ni qit,+
ni qit,? s
+
?
+
t
t
i:s?I
i:s?I
q
q
i
i
i
i
but does not provide a convergence proof. For a fixed-point update, we expect that at the
optimum, the multiplicative factor in (9) is one. However, unlike (5), when the factor is one,
(9) does not relate to ?l(p)/?ps = 0. In fact, a simple example shows that this algorithm
Pk
may never converge. Taking the ?one-against-the rest? approach, if we keep i=1 pti = 1
and assume ni = 1, then ri + ri?P= 1 and the factor in the update rule (9) is
rs + i:i6=s ri?
P
pts + i:i6=s (1?pti )
=
P
k?1+2rs ? k
i=1 ri
.
k?2+2pts
Pk
If the algorithm converges and the factor approaches one, then ps = (1 + 2rs ? i=1 ri )/2
Pk
Pk
but they may not satisfy s=1 ps = 1. Therefore, if in the algorithm we keep i=1 pti = 1
as [13] did, the factor may not approach one and the algorithm does not converge. More
generally, if Ii = {1, . . . , k}, ?i, the algorithm may not converge. As qit = 1, the condition
Pk
that the factor equals one can be written as a linear equation of p. Together with i=1 pi =
1, there is an over-determined linear system (i.e., k + 1 equations and k variables).
6
Experiments on Multi-class Probability Estimates
6.1
Simulated Examples
We consider the same settings in [8, 12] by defining three possible class probabilities:
(a) p1 = 1.5/k, pj = (1 ? p1 )/(k ? 1), j = 2, . . . , k.
(b) k1 = k/2 if k is even, and (k + 1)/2 if k is odd; then p1 = 0.95 ? 1.5/k1 , pi =
(0.95 ? p1 )/(k1 ? 1) for i = 2, . . . , k1 , and pi = 0.05/(k ? k1 ) for i = k1 + 1, . . . , k.
(c) p1 = 0.95 ? 1.5/2, p2 = 0.95 ? p1 , and pi = 0.05/(k ? 2), i = 3, . . . , k.
Classes are competitive in case (a), but only two dominate in case (c). We then generate ri
by adding some noise to qi+ /qi :
q+
ri = min(max(?, qii (1 + 0.1N (0, 1))), 1 ? ?).
Then ri? = 1 ? ri . Here ? = 10?7 is used so that all ri , ri? are positive. We consider the
four encodings used in [1] to generate Ii :
1. ?1vs1?: the pairwise approach (eq. (2)).
2. ?1vsrest?: the ?one-against-the rest? approach in Section 5.1.
3. ?dense?: Ii = {1, . . . , k} for all i. Ii is randomly split to two equally-sized sets Ii+ and
Ii? . [10 log2 k] such splits are generated2 . Following [1], we repeat this procedure 100
times and select the one whose [10 log2 k] splits have the smallest distance.
4. ?sparse?: Ii+ , Ii? are randomly drawn from {1, . . . , k} with E(|Ii+ |) = E(|Ii? |) = k/4.
Then [15 log2 k] such splits are generated. Similar to ?dense,? we repeat the procedure
100 times to find a good coding.
Figure 1 shows averaged accuracy rates over 500 replicates for each of the four methods
when k = 22 , 23 , . . . , 26 . ?1vs1? is good for (a) and (b), but suffers some losses in (c),
where the class probabilities are highly unbalanced. [12] has observed this and proposed
some remedies. ?1vsrest? is quite competitive in all three scenarios. Furthermore, ?dense?
and ?sparse? are less competitive in cases (a) and (b) when k is large. Due to the large
|Ii+ | and |Ii? |, the model is unable to single out a clear winner when probabilities are more
balanced. We also analyze the (relative) mean square error (MSE) in Figure 2:
? k
!
500
k
X
1 X X j
MSE =
(?
p ? pi )2 /
p2i ,
(10)
500 j=1 i=1 i
i=1
? j is the probability estimate obtained in the jth of the 500 replicates. Results of
where p
Figures 2(b) and 2(c) are consistent with those of the accuracy. Note that in Figure 2(a), as
P
? j ) are balanced, ki=1 (?
p (and p
pji ? pi )2 is small. Hence, all approaches have small MSE
though some have poor accuracy.
2
We use [x] to denote the nearest integer value of x.
1
0.8
0.8
0.8
0.6
0.4
0.2
Test Accuracy
1
Test Accuracy
Test Accuracy
1
0.6
0.4
0.2
0
2
3
4
log k
5
0
2
6
0.6
0.4
0.2
3
2
4
log k
5
0
2
6
3
2
(a)
4
log k
5
6
2
(b)
(c)
0.4
0.014
0.35
0.012
0.3
0.01
0.25
0.008
0.4
0.35
0.3
0.2
0.15
0.006
0.004
0.1
0.002
0.05
0
2
0.45
MSE
0.016
MSE
MSE
Figure 1: Accuracy by the four encodings: ?1vs1? (dashed line, square), ?1vsrest? (solid
line, cross), ?dense? (dotted line, circle), ?sparse? (dashdot line, asterisk)
3
4
log2 k
5
6
(a)
0
2
0.25
0.2
0.15
0.1
0.05
3
4
log2 k
(b)
5
6
0
2
3
4
log2 k
5
6
(c)
Figure 2: MSE by the four encodings: legend the same as Figure 1
6.2
Experiments on Real Data
In this section we present experimental results on some real-world multi-class problems.
They have been used in [12], which provides more information about data preparation. Two
problem sizes, 300/500 and 800/1,000 for training/testing, are used. 20 training/testing
splits are generated and the testing error rates are averaged. All data used are available
at http://www.csie.ntu.edu.tw/?cjlin/papers/svmprob/data. We use
the same four ways in Section 6.1 to generate Ii . All of them have |I1 | ? ? ? ? ? |Im |. With
the property that these multi-class problems are reasonably balanced, we set ni = 1 in (8).
Since there are no probability values available for these problems, we compare the accuracy by predicting the label with the largest probability estimate. The purpose here is to
compare the four probability estimates but not to check the difference from existing multiclass classification techniques. We consider support vector machines (SVM) [4] with the
RBF kernel as the binary classifier. An improved version [10] of [11] obtains ri . Full
SVM parameter selection is conducted before testing, although due to space limitation, we
omit details here. The code is modified from LIBSVM [3], a library for support vector
machines. The resulting accuracy is in Table 1 for smaller and larger training/testing sets.
Except ?1vs1,? the other three approaches are quite competitive. These results indicate that
practical problems are more similar to the case of (c) in Section 6.1, where few classes
dominate. This observation is consistent with the findings in [12]. Moreover, ?1vs1? suffers some losses when k is larger (e.g., letter), the same as in Figure 1(c); so for ?1vs1,?
[12] proposed using a quadratic model instead of the Bradley-Terry model.
In terms of the computational time, because the number of binary problems for ?dense? and
?sparse? ([10 log2 k] and [15 log2 k], respectively) is larger than k, and each binary problem
involves many classes of data (all and one half), their training time is longer than ?1vs1?
and ?1vsrest.? ?Dense? is particularly time consuming. Note that though ?1vs1? solves
k(k ? 1)/2 binaries, it is efficient as each binary problem involves only two classes of data.
Table 1: Average of 20 test errors (in percentage) by four encodings (lowest boldfaced)
300 training and 500 testing
800 training and 1,000 testing
Problem k
1vs1 1vsrest dense sparse
1vs1 1vsrest dense sparse
dna
3 10.47
10.33 10.45 10.19
6.21
6.45
6.415
6.345
waveform 3 15.01
15.35 15.66 15.12 13.525 13.635
13.76
13.99
satimage 6 14.22
15.08 14.72
14.8
11.54
11.74 11.865 11.575
segment 7
6.24
6.69
6.62
6.19
3.295
3.605
3.52
3.25
10.89 10.81 11.14
7.78
7.49
7.31 7.575
USPS 10 11.37
MNIST 10 13.84
12.56
13.0 12.29
8.11
7.37
7.59
7.535
35.17 33.86 33.88
21.11 19.685
20.14 19.49
letter
26 39.73
In summary, we propose a generalized Bradley-Terry model which gives individual skill
from group competition results. A useful application to general multi-class probability
estimate is demonstrated.
References
[1] E. L. Allwein, R. E. Schapire, and Y. Singer. Reducing multiclass to binary: a unifying approach
for margin classifiers. Journal of Machine Learning Research, 1:113?141, 2001.
[2] R. A. Bradley and M. Terry. The rank analysis of incomplete block designs: I. the method of
paired comparisons. Biometrika, 39:324?345, 1952.
[3] C.-C. Chang and C.-J. Lin. LIBSVM: a library for support vector machines, 2001. Software
available at http://www.csie.ntu.edu.tw/?cjlin/libsvm.
[4] C. Cortes and V. Vapnik. Support-vector network. Machine Learning, 20:273?297, 1995.
[5] H. A. David. The method of paired comparisons. Oxford University Press, New York, second
edition, 1988.
[6] R. R. Davidson and P. H. Farquhar. A bibliography on the method of paired comparisons.
Biometrics, 32:241?252, 1976.
[7] T. G. Dietterich and G. Bakiri. Solving multiclass learning problems via error-correcting output
codes. Journal of Artificial Intelligence Research, 2:263?286, 1995.
[8] T. Hastie and R. Tibshirani. Classification by pairwise coupling. In M. I. Jordan, M. J. Kearns,
and S. A. Solla, editors, Advances in Neural Information Processing Systems 10. MIT Press,
Cambridge, MA, 1998.
[9] D. R. Hunter. MM algorithms for generalized Bradley-Terry models. The Annals of Statistics,
32:386?408, 2004.
[10] H.-T. Lin, C.-J. Lin, and R. C. Weng. A note on Platt?s probabilistic outputs for support vector
machines. Technical report, Department of Computer Science, National Taiwan University,
2003.
[11] J. Platt. Probabilistic outputs for support vector machines and comparison to regularized likelihood methods. In A. Smola, P. Bartlett, B. Sch?olkopf, and D. Schuurmans, editors, Advances
in Large Margin Classifiers, Cambridge, MA, 2000. MIT Press.
[12] T.-F. Wu, C.-J. Lin, and R. C. Weng. Probability estimates for multi-class classification by pairwise coupling. In S. Thrun, L. Saul, and B. Sch?olkopf, editors, Advances in Neural Information
Processing Systems 16. MIT Press, Cambridge, MA, 2004.
[13] B. Zadrozny. Reducing multiclass to binary by coupling probability estimates. In T. G. Dietterich, S. Becker, and Z. Ghahramani, editors, Advances in Neural Information Processing
Systems 14, pages 1041?1048. MIT Press, Cambridge, MA, 2002.
[14] E. Zermelo. Die berechnung der turnier-ergebnisse als ein maximumproblem der wahrscheinlichkeitsrechnung. Mathematische Zeitschrift, 29:436?460, 1929.
| 2705 |@word version:2 r:10 tried:1 solid:1 initial:2 existing:2 bradley:15 si:2 written:1 dashdot:1 remove:1 update:8 stationary:4 half:1 intelligence:1 ith:2 turnier:1 record:1 provides:1 along:1 prove:4 combine:1 boldfaced:1 introduce:2 pairwise:5 indeed:1 p1:9 multi:14 globally:1 considering:1 becomes:1 bounded:1 moreover:1 lowest:1 r21:1 finding:1 guarantee:1 biometrika:1 classifier:4 platt:2 omit:2 positive:5 before:1 zeitschrift:1 encoding:4 oxford:1 suggests:1 qii:1 averaged:2 unique:1 practical:1 testing:7 block:1 procedure:3 area:2 cannot:1 selection:1 www:2 demonstrated:2 go:3 correcting:5 rule:3 estimator:2 dominate:2 badminton:1 handle:1 searching:1 updated:1 annals:1 pt:34 losing:1 satisfying:1 particularly:1 observed:1 csie:2 rij:5 solved:1 solla:1 decrease:3 balanced:3 solving:5 segment:1 usps:1 train:1 artificial:1 whose:1 heuristic:1 larger:5 solve:4 quite:2 s:1 statistic:3 cov:1 sequence:1 propose:5 date:1 competition:6 normalize:2 olkopf:2 convergence:9 double:2 p:26 r1:5 optimum:1 converges:3 tk:1 coupling:4 nearest:1 odd:1 qt:1 eq:2 solves:1 p2:1 involves:2 implies:3 indicate:1 direction:4 waveform:1 qi0:3 bin:1 ntu:2 im:1 strictly:1 correction:1 mm:1 sufficiently:1 normal:1 smallest:1 purpose:1 ptj:2 label:2 largest:1 weighted:1 mit:4 modified:2 fulfill:1 allwein:1 focus:1 rank:2 likelihood:4 check:1 i1:1 overall:1 issue:1 classification:4 proposes:1 constrained:1 special:3 equal:1 construct:2 never:2 others:1 report:1 hint:1 few:1 randomly:2 national:3 individual:14 n1:1 attempt:1 highly:1 replicates:2 weng:3 biometrics:1 incomplete:1 circle:1 nij:2 instance:1 earlier:1 subset:3 conducted:1 combined:1 st:4 systematic:1 probabilistic:2 together:2 tzu:1 satisfied:2 nm:1 huang:1 zermelo:1 rji:1 potential:1 coding:3 matter:1 satisfy:1 multiplicative:4 closed:1 analyze:1 start:2 competitive:4 minimize:1 square:3 ni:13 accuracy:9 hunter:1 j6:1 suffers:2 against:7 tucker:1 proof:5 stop:1 popular:1 sophisticated:1 back:3 improved:1 though:2 furthermore:1 smola:1 until:2 dietterich:2 remedy:1 hence:4 equality:2 i2:4 game:1 numerator:1 berechnung:1 die:1 svmprob:1 generalized:10 ruby:1 demonstrate:1 winner:1 extend:2 discussed:2 cambridge:4 i6:9 tennis:1 longer:1 add:1 multivariate:1 scenario:1 binary:9 der:2 minimum:1 additional:1 greater:1 converge:3 dashed:1 ii:71 full:1 rj:3 reduces:1 maximumproblem:1 technical:1 cross:1 lin:5 mle:1 equally:1 paired:8 qi:14 prediction:1 denominator:3 iteration:2 kernel:1 sch:2 rest:7 unlike:1 strict:2 subject:3 legend:1 mod:2 jordan:1 integer:1 split:5 viability:1 easy:1 ergebnisse:1 hastie:1 multiclass:4 whether:1 motivated:1 bartlett:1 becker:1 effort:1 rsi:4 york:1 useful:3 generally:1 clear:1 involve:1 dna:1 generate:3 http:2 schapire:1 percentage:1 r12:1 dotted:1 disjoint:2 tibshirani:1 broadly:1 mathematische:1 affected:1 group:3 four:7 drawn:1 pj:21 libsvm:3 letter:2 chih:1 reasonable:1 wu:1 scaling:1 ki:1 convergent:3 quadratic:1 constraint:3 ri:44 software:1 bibliography:1 min:3 department:3 poor:1 remain:1 smaller:2 pti:3 tw:2 restricted:1 errorcorrecting:1 equation:2 vs1:10 discus:3 cjlin:2 needed:1 singer:1 available:5 apply:1 pji:1 original:2 ensure:2 log2:8 unifying:1 qit:13 taipei:2 k1:6 ghahramani:1 bakiri:1 already:1 rt:1 distance:1 unable:1 simulated:1 thrun:1 considers:1 induction:1 taiwan:4 code:6 index:1 kk:1 minimizing:1 relate:1 farquhar:1 negative:4 design:1 observation:1 dispersion:1 finite:1 descent:2 beat:6 zadrozny:1 defining:2 extended:2 team:5 david:1 wahrscheinlichkeitsrechnung:1 address:1 max:1 terry:15 difficulty:1 regularized:1 predicting:2 representing:1 library:2 review:1 asymptotic:3 relative:1 loss:2 expect:1 limitation:2 var:1 asterisk:1 consistent:2 editor:4 pi:13 summary:1 repeat:4 jth:2 saul:1 taking:1 barrier:1 sparse:6 world:1 commonly:1 approximate:4 skill:7 obtains:2 compact:1 keep:2 kkt:2 p0j:3 consuming:1 davidson:1 subsequence:1 iterative:4 sk:1 table:2 reasonably:1 schuurmans:1 mse:7 did:2 pk:12 dense:8 whole:1 noise:1 edition:1 p2i:1 wherep:1 ein:1 winning:1 theorem:6 rk:5 formula:1 jen:1 r2:2 svm:2 cortes:1 mnist:1 vapnik:1 adding:1 margin:2 expressed:1 sport:1 chang:1 satisfies:4 ma:4 sized:1 rbf:1 satimage:1 infinite:1 determined:1 except:1 reducing:2 lemma:2 kearns:1 total:2 kuo:1 experimental:1 select:1 support:6 unbalanced:1 preparation:1 |
1,879 | 2,706 | Making Latin Manuscripts Searchable using
gHMM?s
Jaety Edwards Yee Whye Teh David Forsyth Roger Bock Michael Maire
{jaety,ywteh,daf,bock,mmaire}@cs.berkeley.edu
Grace Vesom
Department of Computer Science
UC Berkeley
Berkeley, CA 94720
Abstract
We describe a method that can make a scanned, handwritten mediaeval
latin manuscript accessible to full text search. A generalized HMM is
fitted, using transcribed latin to obtain a transition model and one example each of 22 letters to obtain an emission model. We show results for
unigram, bigram and trigram models. Our method transcribes 25 pages
of a manuscript of Terence with fair accuracy (75% of letters correctly
transcribed). Search results are very strong; we use examples of variant spellings to demonstrate that the search respects the ink of the document. Furthermore, our model produces fair searches on a document
from which we obtained no training data.
1. Intoduction
There are many large corpora of handwritten scanned documents, and their number is growing rapidly. Collections range from the complete works of Mark Twain to thousands of
pages of zoological notes spanning two centuries. Large scale analyses of such corpora
is currently very difficult, because handwriting recognition works poorly. Recently, Rath
and Manmatha have demonstrated that one can use small bodies of aligned material as
supervised data to train a word spotting mechanism [7]. The result can make scanned
handwritten documents searchable.
Current techniques assume a closed vocabulary ? one can search only for words in the
training set ? and search for instances of whole words. This approach is particularly
unattractive for an inflected language, because individual words can take so many forms
that one is unlikely to see all in the training set. Furthermore, one would like the method
used to require very little aligned training data, so that it is possible to process documents
written by different scribes with little overhead. Mediaeval Latin manuscripts are a natural
first corpus for studying this problem, because there are many scanned manuscripts and
because the handwriting is relatively regular. We expect the primary user need to be search
over a large body of documents ? to allow comparisons between documents ? rather
than transcription of a particular document (which is usually relatively easy to do by hand).
Desirable features for a system are: First, that it use little or no aligned training data (an
ideal, which we believe may be attainable, is an unsupervised learning system). Second,
that one can search the document for an arbitrary string (rather than, say, only complete
words that appear in the training data). This would allow a user to determine whether a
document contains curious or distinctive spellings, for example (figure 7).
We show that, using a statistical model based on a generalized HMM, we can search a medieval manuscript with considerable accuracy, using only one instance each of each letter
in the manuscript to train the method (22 instances in total; Latin has no j, k, w, or z). Furthermore, our method allows fairly accurate transcription of the manuscript. We train our
system on 22 glyphs taken from a a 12th century latin manuscript of Terence?s Comedies
(obtained from a repository of over 80 scanned medieval works maintained by Oxford University [1]). We evaluate searches using a considerable portion of this manuscript aligned
by hand; we then show that fair search results are available on a different manuscript (MS.
Auct. D. 2. 16, Latin Gospels with beast-headed evangelist portraits made at Landvennec,
Brittany, late 9th or early 10th century, from [1]) without change of letter templates.
1.1. Previous Work
Handwriting recognition is a traditional problem, too well studied to review in detail here
(see [6]). Typically, online handwriting recognition (where strokes can be recorded) works
better than offline handwriting recognition. Handwritten digits can now be recognized
with high accuracy [2, 5]. Handwritten amounts can be read with fair accuracy, which
is significantly improved if one segments the amount into digits at the same time as one
recognizes it [4, 5]. Recently several authors have proposed new techniques for search
and translation in this unrestricted setting. Manmatha et al [7] introduce the technique of
?word spotting,? which segments text into word images, rectifies the word images, and
then uses an aligned training set to learn correspondences between rectified word images
and strings. The method is not suitable for a heavily inflected language, because words
take so many forms. In an inflected language, the natural unit to match to is a subset of a
word, rather than a whole word, implying that one should segment the text into blocks ?
which may be smaller than words ? while recognizing. Vinciarelli et al [8] introduce a
method for line by line recognition based around an HMM and quite similar to techniques
used in the speech recognition community. Their method uses a window that slides along
the text to obtain features; this has the difficulty that the same window is in some places too
small (and so uninformative) and in others too big (and so spans more than one letter, and is
confusing). Their method requires a substantial body of aligned training data, which makes
it impractical for our applications. Close in spirit to our work is the approach to machine
translation of Koehn and Knight [3]. They demonstrate that the statistics of unaligned
corpora may provide as powerful constraints for training models as aligned bitexts.
2. The Model
Our models for both search and transcription are based on the generalized HMM and differ
only in their choice of transition model. In an HMM, each hidden node ct emits a single
evidence node xt . In a generalized HMM, we allow each ct to emit a series of x?s whose
length is itself a random variable. In our model, the hidden nodes correspond to letters
and each xt is a single column of pixels. Allowing letters to emit sets of columns lets us
accomodate letter templates of variable width. In particular, this means that we can unify
segmenting ink into letters and recognizing blocks of ink; figure 3 shows an example of
how useful this is.
2.1. Generating a line of text
Our hidden state consists of a character label c, width w and vertical position y. The
statespace of c contains the characters ?a?-?z?, a space ? ?, and a special end state ?. Let T c
be the template associated with character c, Tch , Tcw be respectively the height and width
of that template, and m be the height of the image.
Figure 1: Left, a full page of our manuscript, a 12?th century manuscript of Terence?s
Comedies obtained from [1]. Top right, a set of lines from a page from that document and
bottom right, some words in higher resolution. Note: (a) the richness of page layout; (b)
the clear spacing of the lines; (c) the relatively regular handwriting.
Figure 2: Left, the 22 instances, one per letter, used to train our emission model. These
templates are extracted by hand from the Terence document. Right, the five image channels
for a single letter.
Beginning at image column 1 (and assuming a dummy space before the first character),
?
?
?
?
choose character c ? p(c|c?1...?n ) (an n-gram letter model)
choose length w ? Uniform(Tcw ? k, Tcw + k) (for some small k)
choose vertical position y ? Uniform(1, m ? Tch )
z,y and Tch now define a bounding box b of pixels. Let i and j be indexed from
the top left of that bounding box.
? draw pixel (i, j) ? N (Tcij , ?cij ) for each pixel in b
? draw all pixels above and below b from background gaussian N (?0 , ?0 )
(See 2.2 for greater detail on pixel emission model)
? move to column w + 1 and repeat until we enter the end state ?.
Inference on a gHMM is a relatively straighforward business of dynamic programming. We have used unigram, bigram and trigram models, with each transition
model fitted using an electronic version of Caesar?s Gallic Wars, obtained from
http://www.thelatinlibrary.com. We do not believe that the choice of author
should significantly affect the fitted transition model ? which is at the level of characters
? but have not experimented with this point. The important matter is the emission model.
2.2. The Emission Model
Our emission model is as follows: Given the character c and width w, we generate a template of the required length. Each pixel in this template becomes the mean of a gaussian
which generates the corresponding pixel in the image. This template has a separate mean
image for each pixel channel. The channels are assumed independent given the means.
We train the model by cutting out by hand a single instance of each letter from our corpus
(figure 2). This forms the central portion of the template. Pixels above and below this
Model
Perfect transcription
unigram
bigram
trigram
matching chars
21019
14603
15572
15788
substitutions
0
5487
4597
4410
insertions
0
534
541
507
deletions
0
773
718
695
Table 1: Edit distance between our transcribed Terence and the editor?s version. Note the
trigram model produces significantly fewer letter errors than the unigram model, but that
the error rate is still a substantial 25%.
central box are generated from a single gaussian used to model background pixels (basically
white pixels). We add a third variable yt to our hidden state indicating the vertical position
of the central box. However, since we are uninterested in actually recovering this variable,
during inference we sum it out of the model. The width of a character is constrained to be
close to the width (tw ) of our hand cut example by setting p(w|c) = 0 for w < tw ? k and
w > tw + k. Here k is a small, user defined integer. Within this range, p(w|c) is distributed
uniformly, larger templates are created by appending pixels from the background model to
the template and smaller ones by simply removing the right k-most columns of the hand
cut example.
For features, we generate five image representations, shown in figure 2. The first is a grayscale version of the original color image. The second and third are generated by convolving
the grayscale image with a vertical derivative of gaussian filter, separating the positive
and negative components of this response, and smoothing each of these gradient images
separately. The fourth and fifth are generated similarly but with a horizontal derivative of
gaussian filter. We have experimented with different weightings of these 5 channels. In
practice we use the gray scale channel and the horizontal gradient channels. We emphasize
the horizontal pieces since these seem the more discriminative.
2.3. Transcription
For transcription, we model letters as coming from an n-gram language model, with no
dependencies between words. Thus, the probability of a letter depends on the k letters
before it, where k = n unless this would cross a word boundary in which case the history
terminates at this boundary. We chose not to model word to word transition probabilities
since, unlike in English, word order in Latin is highly arbitrary. This transition model is
fit from a corpus of ascii encoded latin. We have experimented with unigram (i.e. uniform
transition probabilities), bigram and trigram letter models. We can perform transcription by
fitting the maximum likelihood path through any given line. Some results of this technique
are shown in figure 3.
2.4. Search
For search, we rank lines by the probability that they contain our search word. We set up
a finite state machine like that in figure 4. In this figure, ?bg? represents our background
model for that portion of the line not generated by our search word. We can use any of
the n-gram letter models described for transcription as the transition model for ?bg?. The
probability that the line contains the search word is the probability that this FSM takes
path 1. We use this FSM as the transition model for our gHMM, and output the posterior
probability of the two arrows leading into the end state. 1 and 2 are user defined weights,
but in practice the algorithm does not appear to be particular sensitive to the choice of these
parameters. The results presented here use the unigram model.
Editorial translation Orator ad vos venio ornatu prologi:
unigram
b u rt o r
a d u o s
u em
o
o r n a t u
p r o l o g r
b u rt o r
a d v o s
v em
o
o r u a t u
p r o l o g r
fo r a t o r
a d v o s
v en i o
o r n a t u
p r o l o g i
bigram
trigram
Figure 3: We transcribe the text by finding the maximum likelihood path through the
gHMM. The top line shows the standard version of the line (obtained by consensus among
editors who have consulted various manuscripts; we obtained this information in electronic form from http://www.thelatinlibrary.com). Below, we show the line
as segmented and transcribed by unigram, bigram and trigram models; the unigram and
bigram models transcribe one word as ?vemo?, but the stronger trigram model forces the
two letters to be segmented and correctly transcribes the word as ?venio?, illustrating the
considerable benefit to be obtained by segmenting only at recognition time.
1 ? ?1
Path 1
1 ? ?2
a
b
bg
?1
?
bg
Path 2
?2
Figure 4: The finite state machine to search for the word ?ab.? ?bg? is a place holder for the
larger finite state machine defined by our language model?s transition matrix.
3. Results
Figure 1 shows a page from our collection. This is a scanned 12th century manuscript
of Terence?s Comedies, obtained from the collection at [1]. In preprocessing, we extract
individual lines of text by rotating the image to various degrees and projecting the sum of
the pixel values onto the y-axis. We choose the orientation whose projection vector has the
lowest entropy, and then segment lines by cutting at minima of this projection.
Transcription is not our primary task, but methods that produce good transcriptions are
going to support good searches. The gHMM can produce a surprisingly good transcription,
given how little training data is used to train the emission model. We aligned an editors
version of Terence with 25 pages from the manuscript by hand, and computed the edit distance between the transcribed text and the aligned text; as table 1 indicates, approximately
75% of letters are read correctly.
Search results are strong. We show results for two documents. The first set of results
refers to the edition of Terence?s Comedies, from which we took the 22 letter instances.
In particular, for any given search term, our process ranks the complete set of lines. We
used a hand alignment of the manuscript to determine which lines contained each term;
figure 5 shows an overview of searches performed using every word that appears in the
50
100
150
200
250
300
350
400
450
500
550
Figure 5: Our search ranks 587 manuscript lines, with higher ranking lines more likely
to contain the relevant term. This figure shows complete search results for each term that
appears more than three times in the 587 lines. Each row represents the ranked search
results for a term, and a black mark appears if the search term is actually in the line; a
successful search will therefore appear as a row which is wholly dark to the left, and then
wholly light. All 587 lines are represented. More common terms are represented by lower
rows. More detailed results appear in figure 5 and figure 6; this summary figure suggests
almost all searches are highly successful.
document more than three times, in particular, showing which of the ranked set of lines
actually contained the search term. For almost every search, the term appears mainly in
the lines with higher rank. Figure 6 contains more detailed information for a smaller set of
words. We do not score the position of a word in a line (for practical reasons).
Figure 7 demonstrates (a) that our search respects the ink of the document and (b) that for
the Terence document, word positions are accurately estimated. The spelling of mediaeval
documents is typically cleaned up by editors; in our manuscript, the scribe reliably spells
?michi? for the standard ?mihi?. A search on ?michi? produces many instances; a search
on ?mihi? produces none, because the ink doesn?t have any. Notice this phenomenon also
in the bottom right line of figure 7, the scribe writes ?habet, ut consumat nunc cum nichil
obsint doli? and the editor gives ?habet, ut consumat nunc quom nil obsint doli.? Figure 8
shows that searches on short strings produce many words containing that string as one
would wish.
4. Discussion
We have shown that it is possible to make at least some handwritten mediaeval manuscripts
accessible to full text search, without requiring an aligned text or much supervisory data.
Our documents have very regular letters, and letter frequencies ? which can be obtained
from transcribed Latin ? appear to provide so powerful a cue that relatively little detailed
information about letter shapes is required. Linking letter segmentation and recognition
has thoroughly beneficial effects. This suggests that the pool of manuscripts that can be
made accessible in this way is large. In particular, we have used our method, trained on
22 instances of letters from one document, to search another document. Figure 9 shows
the results from two searches of our second document (MS. Auct. D. 2. 16, Latin Gospels
with beast-headed evangelist portraits made at Landvennec, Brittany, late 9th or early 10th
century, from [1]). No information from this document was used in training at all; but letter
1tu
arbitror
pater
etiam
nisi
factum
primum
siet
vero
illi
inter
hic
michi
ibi
qui
tu
ibi
michi
0.9
0.8
0.7 qui
hic
0.6
inter
0.5
illi
0.4
siet
0.3
vero
0.2
nisi
0.1
50
100
150
200
250
300
350
400
450
500
550
0
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
Figure 6: On the left, search results for selected words (indicated on the leftmost column).
Each row represents the ranked search results for a term, and a black mark appears if
the search term is actually in the line; a successful search will therefore appear as a row
which is wholly dark to the left, and then wholly light. Note only the top 300 results are
represented, and that lines containing the search term are almost always at or close to the
top of the search results (black marks to the left). On the right, we plot precision against
recall for a set of different words by taking the top 10, 20, ... lines returned from the search,
and checking them against the aligned manuscript. Note that, once all cases have been
found, if the size of the pool is increased the precision will fall with 100% recall; many
words work well, with most of the first 20 or so lines returned containing the search term.
shapes are sufficiently well shared that the search is still useful.
All this suggests that one might be able to use EM to link three processes: one that clusters
to determine letter shapes; one that segments letters; and one that imposes a language
model. Such a system might be able to make handwritten Latin searchable with no training
data.
References
[1] Early Manuscripts at Oxford University.
Bodleian library ms. auct. f. 2.13.
http://image.ox.ac.uk/.
[2] Serge Belongie, Jitendra Malik, and Jan Puzicha. Shape matching and object recognition using shape contexts. IEEE T. Pattern Analysis and Machine Intelligence,
24(4):509?522, 2002.
[3] Philipp Koehn and Kevin Knight. Estimating word translation probabilities from unrelated monolingual corpora. In Proc. of the 17th National Conf. on AI, pages 711?715.
AAAI Press / The MIT Press, 2000.
[4] Y. LeCun, L. Bottou, and Y. Bengio. Reading checks with graph transformer networks.
In International Conference on Acoustics, Speech, and Signal Processing, volume 1,
pages 151?154, Munich, 1997. IEEE.
[5] Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to
document recognition. Proceedings of the IEEE, 86(11):2278?2324, 1998.
[6] R. Plamondon and S.N. Srihari. Online and off-line handwriting recognition: a comprehensive survey. IEEE Transactions on Pattern Analysis and Machine Intelligence,
22(1):63?84, 2000.
[7] T. M. Rath and R. Manmatha. Word image matching using dynamic time warping.
In Proc. of the Conf. on Computer Vision and Pattern Recognition (CVPR), volume 2,
pages 521?527, 2003.
[8] Alessandro Vinciarelli, Samy Bengio, and Horst Bunke. Offline recognition of unconstrained handwritten texts using hmms and statistical language models. IEEE Trans.
Pattern Anal. Mach. Intell., 26(6):709?720, 2004.
michi: Spe incerta certum mihi laborem sustuli,
mihi: Faciuntne intellegendo ut nil intellegant?
michi: Nonnumquam conlacrumabat. placuit tum id mihi.
mihi: Placuit: despondi. hic nuptiis dictust dies.
michi: Sto exspectans siquid mi imperent. venit una, "heus tu" inquit "Dore,
mihi: Meam ne tangam? CH. Prohibebo inquam. GN. Audin tu? hic furti se adligat:
mihi: Habet, ut consumat nunc quom nil obsint doli;
michi: Quando nec gnatu? neque hic mi quicquam obtemperant,
Figure 7: The handwritten text does not fully correspond to the transcribed version; for
example, scribes commonly write ?michi? for the standard ?mihi?. Our search process
reflects the ink fairly faithfully, however. Left the first four lines returned for a search on
the string ?michi?; right the first four lines returned for a search on the string ?mihi?,
which does not appear in the document. Note that our search process can offer scholars
access to the ink in a particular document, useful for studying variations in transcription,
etc.
tu: Quid te futurum censes quem adsidue exedent?
tu: Quae ibi aderant forte unam aspicio adulescentulam
Figure 8: Searches on short strings produce substrings of words as well as words (we show
the first two lines returned from a search for ?tu?).
sunt
interrogaverunt
Figure 9: The first six lines returned from the second manuscript, (MS. Auct. D. 2. 16, Latin
Gospels with beast-headed evangelist portraits made at Landvennec, Brittany, late 9th or
early 10th century, from [1]), in response to the queries ?interrogeraverunt? (left; lines
three and six contain the word, which is localized largely correctly) and ?sunt? (right;
lines one and four contain the word). We do not have aligned text, so cannot measure the
recall and precision for searches on this document. The recall and precision are clearly
not as good as those for the Terence document, the search is reasonably satisfactory, given
that no training information from this document was available.
| 2706 |@word repository:1 version:6 illustrating:1 bigram:7 stronger:1 attainable:1 manmatha:3 contains:4 series:1 substitution:1 score:1 document:28 rath:2 current:1 com:2 written:1 shape:5 plot:1 implying:1 cue:1 fewer:1 selected:1 intelligence:2 beginning:1 short:2 zoological:1 node:3 philipp:1 five:2 height:2 along:1 consists:1 overhead:1 fitting:1 headed:3 introduce:2 inter:2 growing:1 little:5 window:2 becomes:1 estimating:1 unrelated:1 lowest:1 hic:5 string:7 finding:1 impractical:1 berkeley:3 every:2 transcribes:2 demonstrates:1 uk:1 unit:1 appear:7 segmenting:2 before:2 positive:1 mach:1 oxford:2 id:1 path:5 approximately:1 black:3 chose:1 ibi:3 might:2 studied:1 suggests:3 hmms:1 range:2 practical:1 lecun:2 practice:2 block:2 writes:1 digit:2 maire:1 jan:1 wholly:4 significantly:3 matching:3 projection:2 word:39 regular:3 refers:1 onto:1 close:3 cannot:1 context:1 transformer:1 yee:1 www:2 demonstrated:1 yt:1 layout:1 survey:1 resolution:1 unify:1 century:7 variation:1 heavily:1 user:4 programming:1 us:2 samy:1 recognition:13 particularly:1 cut:2 bottom:2 thousand:1 richness:1 knight:2 substantial:2 alessandro:1 insertion:1 dynamic:2 trained:1 segment:5 brittany:3 distinctive:1 various:2 represented:3 train:6 describe:1 query:1 kevin:1 quite:1 whose:2 koehn:2 larger:2 encoded:1 say:1 cvpr:1 statistic:1 itself:1 online:2 took:1 unaligned:1 coming:1 tu:7 aligned:12 relevant:1 rapidly:1 poorly:1 cluster:1 produce:8 generating:1 perfect:1 object:1 ac:1 strong:2 una:1 recovering:1 c:1 edward:1 differ:1 filter:2 char:1 material:1 require:1 scholar:1 around:1 sufficiently:1 trigram:8 early:4 proc:2 label:1 currently:1 sensitive:1 edit:2 faithfully:1 reflects:1 mit:1 clearly:1 gaussian:5 always:1 rather:3 bunke:1 emission:7 rank:4 likelihood:2 indicates:1 mainly:1 check:1 twain:1 inference:2 unlikely:1 typically:2 hidden:4 going:1 pixel:14 among:1 orientation:1 constrained:1 special:1 fairly:2 uc:1 smoothing:1 once:1 represents:3 unsupervised:1 caesar:1 others:1 national:1 comprehensive:1 individual:2 intell:1 ab:1 highly:2 rectifies:1 alignment:1 light:2 accurate:1 emit:2 unam:1 fsm:2 unless:1 indexed:1 rotating:1 fitted:3 instance:8 portrait:3 column:6 increased:1 gn:1 subset:1 uniform:3 recognizing:2 successful:3 nuptiis:1 too:3 dependency:1 thoroughly:1 international:1 accessible:3 off:1 terence:10 michael:1 pool:2 central:3 recorded:1 aaai:1 containing:3 choose:4 transcribed:7 conf:2 convolving:1 derivative:2 leading:1 matter:1 forsyth:1 jitendra:1 ranking:1 depends:1 bg:5 piece:1 ad:1 performed:1 closed:1 portion:3 accuracy:4 holder:1 who:1 largely:1 correspond:2 serge:1 handwritten:9 accurately:1 basically:1 none:1 substring:1 rectified:1 history:1 stroke:1 fo:1 against:2 frequency:1 spe:1 associated:1 mi:2 handwriting:7 emits:1 recall:4 color:1 ut:4 segmentation:1 actually:4 manuscript:24 appears:5 tum:1 higher:3 supervised:1 response:2 improved:1 box:4 ox:1 furthermore:3 roger:1 until:1 hand:8 horizontal:3 indicated:1 gray:1 believe:2 supervisory:1 glyph:1 effect:1 contain:4 requiring:1 spell:1 read:2 satisfactory:1 white:1 during:1 width:6 forte:1 maintained:1 m:4 generalized:4 whye:1 leftmost:1 complete:4 demonstrate:2 image:15 recently:2 common:1 overview:1 volume:2 linking:1 enter:1 ai:1 unconstrained:1 similarly:1 vos:1 language:7 access:1 etc:1 add:1 posterior:1 minimum:1 unrestricted:1 greater:1 recognized:1 determine:3 signal:1 full:3 desirable:1 vinciarelli:2 segmented:2 match:1 cross:1 offer:1 variant:1 vision:1 editorial:1 background:4 uninformative:1 separately:1 spacing:1 unlike:1 searchable:3 spirit:1 seem:1 integer:1 curious:1 ideal:1 latin:13 bengio:3 easy:1 affect:1 fit:1 haffner:1 whether:1 six:2 war:1 quid:1 returned:6 speech:2 useful:3 clear:1 detailed:3 se:1 amount:2 slide:1 dark:2 http:3 generate:2 notice:1 estimated:1 correctly:4 per:1 dummy:1 write:1 four:3 inflected:3 jaety:2 graph:1 sum:2 letter:30 powerful:2 fourth:1 place:2 almost:3 electronic:2 draw:2 confusing:1 dy:1 qui:2 ct:2 correspondence:1 scanned:6 constraint:1 ywteh:1 generates:1 span:1 relatively:5 department:1 munich:1 smaller:3 terminates:1 em:3 beast:3 character:8 beneficial:1 tw:3 making:1 projecting:1 taken:1 mechanism:1 end:3 studying:2 available:2 appending:1 original:1 top:6 recognizes:1 auct:4 ink:7 warping:1 move:1 malik:1 primary:2 rt:2 spelling:3 grace:1 traditional:1 gradient:3 distance:2 separate:1 link:1 separating:1 hmm:6 evaluate:1 consensus:1 spanning:1 reason:1 assuming:1 length:3 difficult:1 cij:1 negative:1 anal:1 reliably:1 perform:1 teh:1 allowing:1 vertical:4 finite:3 consulted:1 arbitrary:2 community:1 david:1 required:2 cleaned:1 acoustic:1 comedy:4 deletion:1 trans:1 able:2 spotting:2 usually:1 below:3 pattern:4 reading:1 suitable:1 difficulty:1 natural:2 business:1 force:1 ranked:3 library:1 ne:1 axis:1 created:1 extract:1 text:14 review:1 checking:1 fully:1 expect:1 monolingual:1 bock:2 localized:1 straighforward:1 degree:1 imposes:1 editor:5 daf:1 ascii:1 translation:4 row:5 summary:1 repeat:1 surprisingly:1 english:1 offline:2 allow:3 fall:1 template:11 taking:1 inquam:1 fifth:1 ghmm:5 distributed:1 boundary:2 benefit:1 vocabulary:1 transition:10 gram:3 doesn:1 author:2 collection:3 made:4 preprocessing:1 horst:1 commonly:1 transaction:1 emphasize:1 cutting:2 transcription:12 corpus:7 scribe:4 assumed:1 belongie:1 bodleian:1 discriminative:1 grayscale:2 search:55 table:2 learn:1 channel:6 reasonably:1 ca:1 bottou:2 cum:1 arrow:1 whole:2 big:1 bounding:2 edition:1 fair:4 body:3 en:1 sto:1 precision:4 illi:2 position:5 wish:1 late:3 third:2 weighting:1 removing:1 unigram:9 xt:2 showing:1 experimented:3 unattractive:1 evidence:1 nec:1 te:1 accomodate:1 entropy:1 simply:1 likely:1 srihari:1 tch:3 contained:2 ch:1 extracted:1 tcw:3 transcribe:2 shared:1 considerable:3 change:1 plamondon:1 uniformly:1 total:1 nil:3 indicating:1 puzicha:1 mark:4 support:1 statespace:1 phenomenon:1 |
1,880 | 2,707 | Neural network computation by
in vitro transcriptional circuits
Jongmin Kim1 , John J. Hopfield3 , Erik Winfree2
Biology , CNS and Computer Science2 , California Institute of Technology.
Molecular Biology3 , Princeton University.
{jongmin,winfree}@dna.caltech.edu, [email protected]
1
Abstract
The structural similarity of neural networks and genetic regulatory networks to digital circuits, and hence to each other, was noted from the
very beginning of their study [1, 2]. In this work, we propose a simple
biochemical system whose architecture mimics that of genetic regulation and whose components allow for in vitro implementation of arbitrary circuits. We use only two enzymes in addition to DNA and RNA
molecules: RNA polymerase (RNAP) and ribonuclease (RNase). We
develop a rate equation for in vitro transcriptional networks, and derive a correspondence with general neural network rate equations [3].
As proof-of-principle demonstrations, an associative memory task and a
feedforward network computation are shown by simulation. A difference
between the neural network and biochemical models is also highlighted:
global coupling of rate equations through enzyme saturation can lead
to global feedback regulation, thus allowing a simple network without
explicit mutual inhibition to perform the winner-take-all computation.
Thus, the full complexity of the cell is not necessary for biochemical
computation: a wide range of functional behaviors can be achieved with
a small set of biochemical components.
1
Introduction
Biological organisms possess an enormous repertoire of genetic responses to everchanging combinations of cellular and environmental signals. Characterizing and decoding the
connectivity of the genetic regulatory networks that govern these responses is a major challenge of the post-genome era [4]. Understanding the operation of biological networks is intricately intertwined with the ability to create sophisticated biochemical networks de novo.
Recent work developing synthetic genetic regulatory networks has focused on engineered
circuits in bacteria wherein protein signals are produced and degraded [5, 6]. Although
remarkable, such network implementations in bacteria have many unknown and uncontrollable parameters.
We propose a biochemical model system ? a simplified analog of genetic regulatory circuits
? that provides well-defined connectivity and uses nucleic acid species as fuel and signals
that control the network. Our goal is to establish an explicit model to guide the laboratory
construction of synthetic biomolecular systems in which every component is known and
RNAP
transcript
RNase
inhibitor
(A)
activator
DNA switch
(B)
Figure 1: (A) The components of an in vitro circuit. The switch template (blue) is shown
with the activator (red) attached. The dotted box indicates the promoter sequence and the
downstream direction. (B) The correspondence between a neural network and an in vitro
biochemical network. Neuron activity corresponds to RNA transcript concentration, while
synaptic connections correspond to DNA switches with specified input and output.
where quantitative predictions can be tested. Only two enzymes are used in addition to synthetic DNA templates: RNA polymerase, which recognizes a specific promoter sequence
in double-stranded DNA and transcribes the downstream DNA to produce an RNA transcript, and ribonuclease, which degrades RNA but not DNA. In this system, RNA transcript
concentrations are taken as signals. Synthetic DNA templates may assume two different
conformations with different transcription efficiency: ON or OFF. Upon interaction with
a RNA transcript of the appropriate sequence, the DNA template switches between different conformations like a gene regulated by transcription factors. The connectivity ? which
RNA transcripts regulate which DNA templates ? is dictated by Watson?Crick base-pairing
rules and is easy to program. The network computation is powered by rNTP that drives the
synthesis of RNA signals by RNAP, while RNase forces transient signals to decay. With
a few assumptions, we find that this stripped-down analog of genetic regulatory networks
is mathematically equivalent to recurrent neural networks, confirming that a wide range of
programmable dynamical behaviors is attainable.
2
Construction of the transcriptional network
The DNA transcriptional switch. The elementary unit of our networks will be a DNA
switch, which serves the role of a gene in a genetic regulatory circuit. The basic requirements for a DNA switch are to have separate input and output domains, to transcribe poorly
by itself [7], and to transcribe efficiently when an activator is bound to it. A possible mechanism of activation is the complementation of an incomplete promoter region, allowing
more favorable binding of RNAP to the DNA template. Figure 1A illustrates our proposed
design for DNA transcriptional switches and circuits. We model a single DNA switch with
the following binding reactions:
A + I
AI
OFF
D +A
DA
ON
ON
DA + I
D + AI
OFF
where D (blue) is a DNA template with an incomplete promoter region, A (red) is an
activator that complements the incomplete promoter region, and I (green) is an inhibitor
complementary to A. Thus, I can bind free A. Furthermore, activator A contains a ?toehold? region [8] that overhangs past the end of D, allowing inhibitor I to strip off A from
the DA complex. D is considered OFF and DA is considered ON, based on their efficiency
as templates for transcription. This set of binding reactions provides a means to choose the
threshold of the sigmoidal activation function, as will be explained later.
RNAP and RNase drive changes in RNA transcript concentration; their activity is modeled
using a first-order approximation for enzyme kinetics. For the moment, we assume that the
input species (activator and inhibitor) are held at constant levels by external control.
By RNA polymerase
By RNase
kp
k
R ?d ?
DA ? DA + R
?kp
D ? D+R
where 0 < ? < 1 due to lack of activation and ? represents the complete degradation of
RNA products by RNase. kd and kp are set by the concentration of enzymes.
In general, a set of chemical reactions obeying mass action have dynamics described by
?
d[Xi ] X Y
=
k?
[Xj ]rj (p?i ? ri? )
dt
j
?
ri?
where k? is the rate constant,
is the stoichiometry of species Xi as a reactant (typically
0 or 1), and p?i is the stoichiometry of Xi as a product in reaction ?. Analysis of our
system is greatly simplified by the assumption that the binding reactions are fast and go
to completion. We define D tot as the sum of free and bound species:D tot = [D] + [DA].
Similarly, I tot = [I]+[AI] and Atot = [A]+[DA]+[AI]. Then, [DA] depends on D tot and
?, where ? = Atot ? I tot . Because I can scavenge A whether the latter is free or bound
to D, A can activate D only when ? > 0. The amount of [DA] is proportional to ? when
0 < ? < D tot , as shown in Figure 2A. It is convenient to represent this nonlinearity using a
piecewise-linear approximation of a sigmoidal function, specifically, ?(x) = |x+1|?|x?1|
.
2
? where
Thus, we can represent [DA] using ? and a rescaled ?: [DA] = 12 Dtot (1 + ?(?)),
? = 2?
?
D tot ? 1 is called the signal activity. At steady-state, kd [R] = kp [DA] + ?kp [D];
thus,
1 kp tot
? + 1 + ?) .
[R] =
D ((1 ? ?)?(?)
2 kd
If we consider the activator concentration as an input and the steady-state transcript concentration as an output, then the (presumed constant) inhibitor concentration, I tot , sets the
threshold, and the function assumes a sigmoidal shape (Fig. 2D). Adjusting the amount of
template, D tot , sets the magnitude of the output signal and the width of the transition region (Fig. 2C). We can adjust the width of the transition region independent of the threshold
such that a step function would be achieved in the limit. Thus, we have a sigmoidal function with an adjustable threshold, without reliance on cooperative binding of transcription
factors as is common in biological systems [9].
Networks of transcriptional switches. The input domain of a DNA switch is upstream
of the promoter region; the output domain is downstream of the promoter region. This
separation of domains allows us to design DNA switches that have any desired connectivity.
[ DA ]
D
? (x)
tot
1
D
tot
?1
?
[R]
1
[R]
x
?1
(A)
(B)
(C)
A
tot
(D)
A
tot
Figure 2: (A) [DA] as a function of ?. (B) The sigmoid ?(x). (C,D) [R] as a function of
Atot for three values of D tot and I tot , respectively.
We assume that distinct signals in the network are represented as distinct RNA sequences
that have negligible crosstalk (undesired binding of two molecules representing different
signals). The set of legitimate binding reactions is as follows:
OFF
A j+ I j
A jI j
D ij+ A j
D ij Aj
ON
D ij + A j I j
OFF
D ij Aj + I j
ON
where Dij is the DNA template that has the jth input domain and ith output domain, the
activator Aj complements the incomplete promoter region of Dij , and the inhibitor Ij is
complementary to Aj . Note that Ij can strip off Aj from the Dij Aj complex, thus imposing
a sharp threshold as before. Again, we assume fast and complete binding reactions.
The set of enzyme reactions for the transcriptional network is as follows:
By RNA polymerase
kp
Dij Aj ? Dij Aj + Ai
By RNase
if sij = 1
?kp
Dij ? Dij + Ai
kp
Dij Aj ? Dij Aj + Ii
?kp
Dij ? Dij + Ii
k
Ij ?d ?
k
if sij = ?1
Aj ?d ?
k
Aj Ij ?d ?
k
Dij Aj ?d Dij
where sij ? {+1, ?1} indicates whether switch ij will produce an activator or an inhibitor.
This notation reflects that the production of Ii is equivalent to the consumption of Ai . The
tot
change of RNA concentrations over time is easy to express with ?i = Atot
i ? Ii :
X
d?i
= ?kd ? ?i + kp
sij ([Dij Aj ] + ?[Dij ]) .
(1)
dt
j
Network equivalence. We show next that the time evolution of this biochemical network
model is equivalent to that of a general Hopfield neural network model [3]:
X
dxi
wij ?(xj ) + ?i .
(2)
= ?xi +
?
dt
j
Equation 1 can be rewritten to use the same nonlinear activation function ? defined earlier.
i
? i = 2?
Let ?
tot ? 1 be a rescaled difference between activator and inhibitor concentrations,
D?i
tot
where DP
?i is the load on Ai , i.e., the total concentration of all switches that bind to A i :
tot
tot
tot
D?i = j Dji
and Dij
= [Dij Aj ] + [Dij ]. Then, we can derive the following rate
? i plays the role of unit i?s activity xi :
equation, where ?
X
tot
tot
X kp
?i
Dij
Dij
kp
1 d?
?
?
= ? ?i +
(1 ? ?)sij tot ?(?j )+
(1 + ?)sij tot ? 1 .
kd dt
kd
D?i
kd
D?i
j
j
(3)
Given the set of constants describing an arbitrary transcriptional network, the constants for
an equivalent neural network can be obtained immediately by comparing Equations 2 and
3. The time constant ? is the inverse of the RNase degradation rate: fast turnover of RNA
molecules leads to fast response of the network. The synaptic weight wij is proportional to
the concentration of switch template ij, attenuated by the load on Ai . However, the threshold ?i is dependent on the weights, perhaps implying a lack of generality. To implement
an arbitrary neural network, we must introduce two new types of switches to the transcriptional network. To achieve arbitrary thresholds, we introduce bias switches D iB which
have no input domain and thus produce outputs constitutively; this adds an adjustable constant to the right hand side of Equation 3. To balance the load on Ai , we add null switches
tot
D0i which bind to Ai but have no output domain; this allows us to ensure that all D?i
are
equal. Consequently, given any neural network with weights wij and thresholds ?i , we can
tot
specify concentrations Dij
such that the biochemical network has identical dynamics, for
some ? .
Michaelis?Menten enzyme reactions. Next, we explore the validity of our assumption
that enzyme kinetics are first-order reactions. A basic but more realistic model is the
Michaelis?Menten mechanism [10], in which the enzyme and substrate bind to form an
enzyme-substrate complex. For example, if E is RNAP,
k+
kcat
E + Dij Aj *
) EDij Aj ? E + Dij Aj + Ii /Ai .
k?
An important ramification of Michaelis?Menten reactions is that there is competition for
the enzyme by the substrates, because the concentration of available enzymes is reduced
as they bind to substrates, leading to saturation when the enzyme concentration is limiting. Using the steady-state assumption for Michaelis?Menten reactions, we establish the
following relations to the rate constants of first-order reactions:
kp =
E tot kcat
?
1 + L KM
? ? kp =
0
E tot kcat
? 0
1 + L KM
kd =
Edtot
kd,cat
?
1 + Ld Kd,M
(4)
where kcat and KM = (k? + kcat )/k+ are the catalytic constant (enzyme?s speed) and
0
Michaelis constant (enzyme?s affinity to target) of RNAP for the ON state switch, k cat
and
0
KM are for the OFF state switch, and kd,cat and Kd,M are the constants of RNase. E tot
P [D A ]
and Edtot are the concentrations of RNAP and RNase, respectively. L = i,j KijM j +
P [Aj ]+[Ij ]+[Aj Ij ]+[Dij Aj ]
P [Dij ]
is the load on RNAP and Ld =
is the load on
0
i,j
i,j KM
Kd,M
RNase (i.e., the total concentration of binding targets divided by the Michaelis constants of
the enzymes), both of which may be time varying. To make the first-order approximation
valid, we must keep L and Ld constant. Introduction of a new type of switch with different
Michaelis constants can make L constant by balancing the load on the enzyme. A scheme
to keep Ld constant is not obvious, so we set reaction conditions such that Ld 1.
3
Example computations by transcriptional networks
Feed-forward networks.
We first consider a feed-forward network to compute
f (x, y, z) = x
?yz + y?z + x. From the Boolean circuit shown in Figure 3A, we can construct
an equivalent neural network. We label units 1 through 6: units 1, 2, 3 correspond to inputs
x, y, z whereas units 4, 5, 6 are computation units. Using the conversion rule discussed
in the network equivalence section, we can calculate the parameters of the transcriptional
network. Under the first-order approximation of Equation 3, the simulation result is exact
(Fig. 3C). For comparison, we also explicitly simulated mass action dynamics for the full
set of chemical equations with the Michaelis?Menten enzyme reactions, using biologically
plausible rate constants and with E tot and Edtot calculated from Equation 4 using estimated
values of L and Ld . The full model performs the correct calculation of f for all eight 3-bit
inputs, although the magnitude of signals is exaggerated due to an underestimate of RNase
load (Fig. 3C).
Associative memories. Figure 4A shows three 4-by-4 patterns to be memorized in a continuous neural network [3]. We chose orthogonal patterns because a 16 neuron network has
limited capacity. Our training algorithm is gradient descent combined with the perceptron
learning rule. After training, the parameters of the neural network are converted to the
parameters of the transcriptional network as previously described. Starting from a random
6
f
1
?1
y
y
z
(B)
z
4
?2 1
1
2
2
1
f
1
0
?2
?1
1
(A)
1
?1
^
?
i
x
x
?4
(C)
0
200
400
600
time(sec)
800
1000
Figure 3: (A,B) A Boolean circuit and a neural network to compute f (x, y, z) = x
?yz+?
y z+
x. (C) The activity of computation units (first-order approximation: solid lines; MichaelisMenten reaction: dotted lines) for x=True=1, y=False=?1, z=True=1.
3
2
^
?
i
1
0
?1
?2
(A)
(B)
?3
0
200
400
time(sec)
600
Figure 4: (A) The three patterns to be memorized. (B) Time-course for the transcriptional
network recovery of the third pattern. (odd columns: blue lines, even columns: red lines)
initial state, a typical response of the transcriptional network (with the first-order approximation of Equation 3) is shown in Figure 4B. Thus, our in vitro transcriptional networks
can support complex sets of stable steady-states.
A winner-take-all network. Instead of trying to compensate for the saturation phenomena
of Michaelis?Menten reactions, we can make use of it for computation. As an example,
consider the winner-take-all computation [11], which is commonly implemented as a neural network with O(N 2 ) mutually inhibitory connections (Fig. 5A), but which can also be
implemented as an electrical circuit with O(N ) interconnections by using a single global
inhibitory feedback gate [12]. In a biochemical system, a limited global resource, such as
RNAP, can act to regulate all the DNA switches and thus similarly produce global inhibition. This effect is exploited by the simple transcriptional network shown in Figure 5B, in
which the output from each DNA switch activates the same DNA switch itself, and mutual
inhibition is achieved by competition for RNAP. Specifically, we have switch templates D ii
with fixed thresholds set by Ii , and Dii produces Ai as its output RNA. With the instant
binding assumption, we then derive the following equation:
dAtot
E tot
kcat
kd,cat tot
E tot
k0
i
=? d
?
Ai +
[Dii Ai ] + cat
[D
]
.
(5)
ii
0
dt
1 + Ld Kd,M
1 + L KM
KM
The production rate of Ai depends on Atot
and on L, while the degradation rate of Ai
i
depends on Atot
and
on
L
,
as
shown
in
Figure
6A. For a winner-take-all network, an ON
d
i
state switch draws more RNAP than an OFF state switch (because of the smaller Michaelis
constant for the ON state). Thus, if the other switches are turned OFF, the load on RNAP
(L) becomes small, leading to faster production of the remaining ON switches. When the
production rate curve and the degradation rate curve have three intersections, bistability is
achieved such that the switches remain ON or OFF, depending on their current state.
Consider n equivalent switches starting with initial activator concentrations above the
threshold, and with the highest concentration at least ? above the rest (as a percentage).
Analysis indicates that a less leaky system (small ?) and sufficient differences in initial
activator concentrations (large ?) can guarantee the existence of a unique winner. Simulations of a 10-switch winner-take-all network confirm this analysis, although we do not see
perfect behavior (Fig. 6B). Figure 6C shows a time-course of a unique winner situation.
Switches get turned OFF one by one whenever the activator level approaches the threshold,
until only one switch remains ON.
?1
?1
0.5
?1
0.5
?1
(A)
1
?1
0.5
0.5
0.5
0.5
1
1
1
?1
1
(B)
1
Figure 5: (A) A 3-unit WTA network with explicit mutual inhibition. (B) An equivalent
biochemical network.
1
25
0.8
20
1/?
0.6
15
0.4
L : high
10
tot
Ii
tot
tot
Ii +D ii
tot
Ai
(B)
2
1.5
1
0.2
5
(A)
3
2.5
i
L : low
dt
30
[A ] / ?M
tot
dAi
0.5
5
10
?(%)
15
0
(C)
0
0
5000
10000
time(sec)
15000
Figure 6: For WTA networks: (A) Production rates (solid lines) for two different L?s,
compared to a linear degradation rate (dotted line). (B) Empirical probability of correct
output as a function of ? and ?. (C) Time-course with ? = 0.33% and ? = 0.04.
Similarly, we can consider a k-WTA network where k winners persist. If we set the parameters appropriately such that k winners are stable but k + 1 winners are unstable, the
simulation result recovered k winners most of the time. Even a single k-WTA gate can
provide impressive computational power [13].
4
Discussion
We have shown that, if we treat transcriptionally controlled DNA switches as synapses
and the concentrations of RNA species as the states of neurons, then the in vitro transcriptional circuit is equivalent to the neural network model and therefore can be programmed
to carry out a wide variety of tasks. The structure of our biochemical networks differs
from that of previous formal models of genetic regulatory circuits [14, 15, 16]. For example, consider the work of [16], which established a connection to the class of Boltzmann
machines. There, the occupancy of regulatory binding sites corresponds to the state of neurons, the weights are set by the cooperative interaction among transcription factors, and the
thresholds are the effective dissociation constants at a binding site. Thus, implementing a
general N -unit neural network requires only O(N ) biochemical species, but up to O(N 2 )
significant binding interactions must be encoded in the molecular sequences. Changing
or tuning a network is therefore non-trivial. In contrast, in our transcriptional networks,
each weight and threshold is represented by the continuously adjustable concentration of a
distinct species, and the introduction or deletion of any node is straightforward.
Each synapse is represented by a DNA switch with a single input?output specification, so
the number of DNA switches grows as O(N 2 ) for a fully recurrent neural network with
N neurons (unlike the circuits of [16]). This constraint may be relieved because, in many
networks of interest, most nodes have a small number of connections [17, 18]. The time
for computation will increase as O(N ) due to finite hybridization rates because, if the total
concentration of all RNA signals is capped, the concentration of any given species will
decrease as 1/N . The weights are capped by the maximum gain of the system, which is the
production rate divided by the degradation rate. Since the time constant of the network is
the inverse of the degradation rate, if we wish to implement a network with large weights,
we must increase the time constant.
We can analyze the cost of computing by considering basic physical chemistry. The energy
consumption is about 20kT (= 10?19 J) per nucleotide incorporated, and 1 bit of informa-
tion is encoded by a sequence containing tens of nucleotides. The encoding energy is large,
since the molecule for each bit must contain specific instructions for connectivity, unlike
spatially arranged digital circuits where a uniform physical signal carrier can be used. Furthermore, many copies (e.g., 1013 for a 1?M signal in 20?l) of a given species must be
produced to change the concentration in a bulk sample. Worse yet, because degradation is
not modulated in the transcriptional network, switching relies on selective change of production rates, thus continually using energy to maintain an ON state. Devising a scheme to
minimize maintenance energy costs, such as in CMOS technology for electrical circuits, is
an important problem.
The theory presented here is meant to serve as a guide for the construction of real biochemical computing networks. Naturally, real systems will deviate considerably from the
idealized model (although perhaps less so than do neural network models from real neurons). For example, hybridization is neither instantaneous nor irreversible, strands can have
undesired conformations and crosstalk, and enzyme reactions depend on the sequence and
are subject to side reactions that generate incomplete products. Some problems, such as hybridization speed and crosstalk, can be reduced by slowing the enzyme reactions and using
proper sequence design [19]. Ultimately, some form of fault tolerance will be necessary at
the circuit level. Restoration of outputs to digital values, achieved by any sufficiently highgain sigmoidal activation function, provides some level of immunity to noise at the gate
level, and attractor dynamics can provide restoration at the network level. A full understanding of fault tolerance in biochemical computing remains an important open question.
Future directions include utilizing the versatility of active RNA molecules (such as aptamers, ribozymes, and riboswitches [20, 21]) for more general chemical input and output,
devising a biochemical learning scheme analogous to neural network training algorithms
[22], and studying the stochastic behavior of the transcriptional network when a very small
number of molecules are involved in small volumes [5].
Acknowledgements. We thank Michael Elowitz, Paul Rothemund, Casimir Wierzynski,
Dan Stick and David Zhang for valuable discussions, and ONR and NSF for funding.
References
[1]
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
[21]
[22]
McCulloch WS, Pitts W, Bull. Math. Biophys. 5 (1943), 115.
Monod J, Jacob F, Cold Spring Harb. Symp. Quant. Biol. 26 (1961), 389-401.
Hopfield JJ, Proc. Nat. Acad. Sci. USA 81 (1984), 3088-3092.
Hasty J, McMillen D, Issacs F, Collins JJ, Nat. Rev. Genet. 2 (2001), 268-279.
Elowitz MB, Leibler S, Nature 403 (2000), 335-338.
Gardner TS, Cantor CR, Collins JJ, Nature 403 (2000), 339-342.
Martin CT, Coleman JE, Biochemistry 26 (1987), 2690-2696.
Yurke B, Mills AP Jr., Genetic Programming and Evolvable Machines 4 (2003), 111-122.
Shea MA, Ackers GK, J. Mol. Biol. 181 (1985), 211-230.
Hammes GG, Thermodynamics and kinetics for the biological sciences, Wiley (2000).
Yuille AL, Gieger D, in The Handbook of Brain Theory and Neural Networks, Arbib MA, ed.,
MIT Press (1995), 1056-1060.
Tank DW, Hopfield JJ, IEEE Trans. on Circuits and Systems 33 (1986), 533-541.
Maass W, Neural Computation 12 (2000), 2519-2535.
Glass L, Kauffman SA, J. Theo. Biol. 39 (1973), 103-129.
Mjolsness E, Sharp DH, Reinitz J, J. Theo. Biol. 152 (1991), 429-453.
Buchler NE, Gerland U, Hwa T, Proc. Nat. Acad. Sci. USA 100 (2003), 5136-5141.
Bray D, Science 301 (2003), 1864-1865.
Reed RD, IEEE Trans. on Neural Networks, 4 (1993), 740-744.
Dirks R, Lin M, Winfree E, Pierce NA, Nucleic Acids Research 32 (2004), 1392-1403.
Lilley DM, Trends Biochem. Sci. 28 (2003), 495-501.
Nudler E, Mironov AS, Trends Biochem. Sci. 29 (2004), 11-17.
Mills AP Jr., Yurke B, Platzman PM, Biosystems 52 (1999), 175-180.
| 2707 |@word open:1 instruction:1 km:7 simulation:4 jacob:1 attainable:1 solid:2 carry:1 ld:7 moment:1 initial:3 contains:1 genetic:10 past:1 reaction:20 current:1 comparing:1 recovered:1 activation:5 yet:1 must:6 john:1 tot:40 realistic:1 confirming:1 shape:1 implying:1 devising:2 jongmin:2 slowing:1 beginning:1 ith:1 coleman:1 provides:3 math:1 node:2 sigmoidal:5 zhang:1 pairing:1 dan:1 symp:1 introduce:2 presumed:1 behavior:4 nor:1 brain:1 considering:1 becomes:1 notation:1 circuit:18 mass:2 fuel:1 null:1 mcculloch:1 guarantee:1 quantitative:1 every:1 act:1 transcribes:1 stick:1 control:2 unit:9 continually:1 before:1 negligible:1 carrier:1 bind:5 treat:1 limit:1 irreversible:1 switching:1 era:1 encoding:1 acad:2 ap:2 chose:1 equivalence:2 limited:2 programmed:1 range:2 unique:2 crosstalk:3 implement:2 differs:1 cold:1 empirical:1 convenient:1 polymerase:4 protein:1 get:1 equivalent:8 go:1 straightforward:1 starting:2 focused:1 recovery:1 immediately:1 legitimate:1 rule:3 mironov:1 utilizing:1 dw:1 analogous:1 limiting:1 construction:3 play:1 target:2 exact:1 substrate:4 programming:1 us:1 trend:2 persist:1 cooperative:2 role:2 electrical:2 calculate:1 region:9 mjolsness:1 decrease:1 rescaled:2 highest:1 valuable:1 govern:1 complexity:1 turnover:1 dynamic:4 ultimately:1 depend:1 ackers:1 serve:1 upon:1 yuille:1 efficiency:2 hopfield:4 k0:1 represented:3 cat:5 distinct:3 fast:4 effective:1 activate:1 kp:15 whose:2 d0i:1 encoded:2 plausible:1 interconnection:1 novo:1 ability:1 highlighted:1 itself:2 associative:2 sequence:8 propose:2 interaction:3 product:3 mb:1 turned:2 ramification:1 poorly:1 achieve:1 competition:2 double:1 requirement:1 produce:5 perfect:1 cmos:1 derive:3 develop:1 coupling:1 recurrent:2 completion:1 depending:1 ij:12 odd:1 conformation:3 transcript:8 sa:1 implemented:2 direction:2 correct:2 stochastic:1 engineered:1 transient:1 memorized:2 dii:2 implementing:1 uncontrollable:1 repertoire:1 biological:4 elementary:1 mathematically:1 kinetics:3 sufficiently:1 considered:2 pitt:1 biochemistry:1 major:1 favorable:1 proc:2 label:1 create:1 reflects:1 mit:1 inhibitor:8 rna:21 activates:1 cr:1 varying:1 catalytic:1 indicates:3 greatly:1 contrast:1 glass:1 dependent:1 biochemical:16 typically:1 w:1 relation:1 wij:3 selective:1 tank:1 among:1 hasty:1 mutual:3 equal:1 construct:1 biology:1 represents:1 identical:1 mimic:1 future:1 piecewise:1 few:1 transcriptionally:1 cns:1 attractor:1 maintain:1 versatility:1 interest:1 mcmillen:1 adjust:1 held:1 kt:1 necessary:2 bacteria:2 nucleotide:2 orthogonal:1 incomplete:5 desired:1 reactant:1 biosystems:1 column:2 earlier:1 boolean:2 bistability:1 restoration:2 bull:1 cost:2 uniform:1 dij:25 dtot:1 synthetic:4 combined:1 considerably:1 intricately:1 off:13 decoding:1 michael:1 synthesis:1 continuously:1 na:1 connectivity:5 again:1 containing:1 choose:1 worse:1 external:1 leading:2 converted:1 de:1 chemistry:1 sec:3 explicitly:1 depends:3 idealized:1 later:1 tion:1 analyze:1 red:3 michaelis:10 minimize:1 hwa:1 degraded:1 acid:2 efficiently:1 dissociation:1 correspond:2 produced:2 drive:2 complementation:1 synapsis:1 evolvable:1 whenever:1 synaptic:2 strip:2 ed:1 underestimate:1 energy:4 involved:1 obvious:1 dm:1 naturally:1 proof:1 dxi:1 gain:1 adjusting:1 sophisticated:1 feed:2 dt:6 response:4 wherein:1 specify:1 synapse:1 arranged:1 box:1 generality:1 furthermore:2 until:1 hand:1 nonlinear:1 biomolecular:1 lack:2 aj:21 perhaps:2 grows:1 usa:2 effect:1 validity:1 contain:1 true:2 evolution:1 hence:1 chemical:3 spatially:1 laboratory:1 leibler:1 maass:1 undesired:2 width:2 noted:1 steady:4 trying:1 gg:1 complete:2 performs:1 instantaneous:1 funding:1 common:1 sigmoid:1 functional:1 dji:1 vitro:7 ji:1 physical:2 winner:11 attached:1 volume:1 analog:2 organism:1 discussed:1 significant:1 imposing:1 ai:18 tuning:1 rd:1 pm:1 similarly:3 nonlinearity:1 menten:6 harb:1 stable:2 specification:1 similarity:1 impressive:1 inhibition:4 base:1 ribonuclease:2 add:2 enzyme:20 biochem:2 recent:1 dictated:1 exaggerated:1 cantor:1 onr:1 watson:1 fault:2 exploited:1 caltech:1 dai:1 signal:14 ii:11 full:4 rj:1 faster:1 calculation:1 compensate:1 lin:1 divided:2 post:1 molecular:2 controlled:1 prediction:1 basic:3 maintenance:1 winfree:2 represent:2 achieved:5 cell:1 addition:2 whereas:1 appropriately:1 rest:1 unlike:2 posse:1 subject:1 hybridization:3 structural:1 feedforward:1 easy:2 switch:37 xj:2 variety:1 architecture:1 arbib:1 quant:1 attenuated:1 genet:1 whether:2 reinitz:1 jj:4 action:2 programmable:1 amount:2 ten:1 dna:27 reduced:2 generate:1 percentage:1 nsf:1 inhibitory:2 dotted:3 estimated:1 aptamers:1 per:1 bulk:1 blue:3 intertwined:1 express:1 reliance:1 threshold:13 enormous:1 changing:1 neither:1 downstream:3 sum:1 inverse:2 separation:1 draw:1 bit:3 bound:3 ct:1 correspondence:2 activity:5 bray:1 constraint:1 relieved:1 ri:2 speed:2 spring:1 martin:1 developing:1 combination:1 kd:15 jr:2 smaller:1 remain:1 wta:4 biologically:1 rev:1 explained:1 sij:6 taken:1 kim1:1 equation:12 mutually:1 previously:1 resource:1 describing:1 remains:2 mechanism:2 serf:1 end:1 studying:1 available:1 operation:1 rewritten:1 eight:1 appropriate:1 regulate:2 gate:3 existence:1 assumes:1 remaining:1 ensure:1 include:1 recognizes:1 instant:1 yz:2 establish:2 question:1 activator:13 degrades:1 concentration:24 transcriptional:20 regulated:1 dp:1 affinity:1 gradient:1 separate:1 thank:1 simulated:1 capacity:1 sci:4 consumption:2 cellular:1 unstable:1 trivial:1 erik:1 modeled:1 reed:1 demonstration:1 balance:1 regulation:2 gk:1 implementation:2 design:3 proper:1 boltzmann:1 unknown:1 perform:1 allowing:3 adjustable:3 conversion:1 nucleic:2 neuron:6 finite:1 descent:1 t:1 situation:1 incorporated:1 dirk:1 arbitrary:4 sharp:2 david:1 complement:2 specified:1 connection:4 immunity:1 california:1 deletion:1 established:1 trans:2 capped:2 dynamical:1 pattern:4 kauffman:1 challenge:1 saturation:3 program:1 green:1 memory:2 power:1 force:1 representing:1 scheme:3 occupancy:1 thermodynamics:1 technology:2 ne:1 gardner:1 deviate:1 understanding:2 acknowledgement:1 powered:1 fully:1 proportional:2 remarkable:1 digital:3 sufficient:1 principle:1 balancing:1 production:7 course:3 free:3 copy:1 jth:1 theo:2 formal:1 guide:2 allow:1 bias:1 institute:1 wide:3 template:12 characterizing:1 stripped:1 side:2 perceptron:1 leaky:1 tolerance:2 feedback:2 calculated:1 curve:2 transition:2 valid:1 genome:1 stranded:1 forward:2 commonly:1 simplified:2 transcription:5 gene:2 keep:2 confirm:1 global:5 active:1 handbook:1 xi:5 continuous:1 regulatory:8 nature:2 molecule:6 mol:1 complex:4 upstream:1 domain:8 da:15 promoter:8 noise:1 paul:1 complementary:2 fig:6 site:2 je:1 wiley:1 explicit:3 obeying:1 wish:1 ib:1 third:1 elowitz:2 down:1 load:8 specific:2 decay:1 false:1 shea:1 magnitude:2 nat:3 pierce:1 illustrates:1 biophys:1 intersection:1 mill:2 explore:1 strand:1 binding:13 corresponds:2 informa:1 environmental:1 relies:1 dh:1 ma:2 transcribe:2 goal:1 consequently:1 crick:1 change:4 specifically:2 typical:1 degradation:8 called:1 specie:9 total:3 support:1 latter:1 modulated:1 meant:1 collins:2 phenomenon:1 princeton:2 tested:1 biol:4 |
1,881 | 2,708 | Who?s in the Picture?
Tamara L. Berg, Alexander C. Berg, Jaety Edwards and D.A. Forsyth
Computer Science Division
U.C. Berkeley
Berkeley, CA 94720
[email protected]
Abstract
The context in which a name appears in a caption provides powerful cues
as to who is depicted in the associated image. We obtain 44,773 face images, using a face detector, from approximately half a million captioned
news images and automatically link names, obtained using a named entity recognizer, with these faces. A simple clustering method can produce fair results. We improve these results significantly by combining
the clustering process with a model of the probability that an individual
is depicted given its context. Once the labeling procedure is over, we
have an accurately labeled set of faces, an appearance model for each
individual depicted, and a natural language model that can produce accurate results on captions in isolation.
1
Introduction
It is a remarkable fact that pictures and their associated annotations are complementary.
This observation has been used to browse museum collections ([1]) and organize large
image collections ([2, 7, 12, 13]). All of these papers use fairly crude ?bag of words?
models, treating words as floating tags and looking at the co-occurrence of image regions
and annotated words. In this paper, we show that significant gains are available by treating
language more carefully.
Our domain is a large dataset of news photographs with associated captions. A face detector
is used to identify potential faces and a named entity recognizer to identify potential names.
Multiple faces and names from one image-caption pair are quite common. The problem is
to find a correspondence between some of the faces and names. As part of the solution
we learn an appearance model for each pictured name and the likelihood of a particular
instance of a name being pictured based on the surrounding words and punctuation.
Face recognition cannot be surveyed reasonably in the space available. Reviews appear
in [6, 10, 11]. Although face recognition is well studied, it does not work very well in
practice [15]. One motivation for our work is to take the large collection of news images
and captions as semi-supervised input and and produce a fully supervised dataset of faces
labeled with names. The resulting dataset exhibits many of the confounding factors that
make real-world face recognition difficult, in particular modes of variation that are not
found in face recognition datasets collected in laboratories. It is important to note that this
task is easier than general face recognition because each face has only a few associated
names.
N1
F1
N1
F1
N2
F2
N2
F2
Null
F1 =
F2 =
N1 = John Furlong,
Null
Null
N1
F1
N1
F1
N2
F2
N2
F2
Null
Null
Null
Null
N1
F1
N1
F1
N2
F2
N2
F2
N2 = Winter Olympics,
Null
Null
President and Chief Operating Officer of the Vancouver,
British Columbia 2010 Bid Corporation John Furlong
(rear) smiles while celebrating with compatriots their
victory in obtaining the 2010 Winter Olympics bid on late
July 2, 2003 in Prague. Vancouver won with 56 votes
against 53 votes for Pyeonchang in the second round of
balloting at an IOC gathering in Prague. REUTERS/Petr
Josek
Null
Null
N1
F1
N2
F2
Null
Null
Null
All possible name/face assignments
Figure 1: Left: A typical data item consisting of a picture and its associated caption.
Center: Detected faces and names for this data item. Right: The set of possible correspondences for this data item. Our model allows each face to be assigned to at most one
name, each name to be assigned to at most one face, and any face or name to be assigned
to Null. Our named entity recognizer occasionally identifies strings that do not refer to
actual people (e.g. ?Winter Olympics?) . These names are assigned low probability under
our model and therefore their assignment to a face is unlikely. EM iterates between computing the expectation of the possible face-name correspondences and updating the face
clusters and language model. Unusually, we can afford to compute all possible face-name
correspondences for a data item since the number of possibilities is small. For this item,
we correctly choose the best matching ?F1 to Null?, ?N2 to Null?, and ?F2 to N1?.
Language: Quite simple common phenomena in captions suggest using a language model.
First, our named entity recognizer occasionally marks incorrect names like ?United Nations?. The context in which these incorrect detections occur suggest that they do not refer
to actual people. Second, name-context pairs can be weighted according to their probability. In a caption such as ?Michael Jackson responds to questioning Thursday, Nov. 14, 2002
in Santa Maria Superior Court in San ta Maria, Calif., during a $21 million lawsuit brought
against him by Marcel Avram for failing to appear at two millennium concerts...?, Michael
Jackson appears in a more favorable context (at the beginning of the caption, followed by
a verb) than Marcel Avram (near the middle of the caption, followed by a preposition).
Our approach combines a simple appearance model using kPCA and LDA, with a language model, based on context. We evaluate both an EM and maximum likelihood clustering and show that incorporating language with appearance produces better results than
using appearance alone. We also show the results of the learned natural language classifier
applied to a set of captions in isolation.
2
Linking a face and language model with EM
A natural way of thinking about name assignment is as a hidden variable problem where
the hidden variables are the correct name-face correspondences for each picture. This suggests using an expectation maximization (EM) procedure. EM iterates between computing
before ? CEO Summit
after ? Martha Stewart
before ? U.S. Joint
after ? Null
before ? U.S. House
before ? Julia Vakulenko
after ? Andrew Fastow after ? Jennifer Capriati
before ? Angelina Jolie
after ? Jon Voight
before ? Ric Pipino
after ? Heidi Klum
before ? Vice President before ? Marcel Avram
Dick Cheney
after ? Michael Jackson
after ? President George W.
before ? U.S. Open
after ? David Nalbandian
before ? al Qaeda
after ? Null
before ? James Bond
after ? Pierce Brosnan
before ? James Ivory
after ? Naomi Watts
Figure 2: Names assigned using our raw clustering procedure (before) and incorporating a
language model (after). Our named entity recognizer occasionally detects incorrect names
(e.g. ?CEO Summit?), but based on context the language model assigns low probabilities
to these names, making their assignment unlikely. When multiple names are detected like
?Julia Vakulenko? and ?Jennifer Capriati?, the probability for each name depends on its
context. The caption for this picture reads ?American Jennifer Capriati returns the ball to
her Ukrainian opponent Julia Vakulenko in Paris during...? The language model prefers to
assign the name ?Jennifer Capriati? because its context (beginning of the caption followed
by a present tense verb) indicates it is more likely to be pictured than ?Julia Vakulenko?
(middle of the caption followed by a preposition). For pictures such the man labeled ?al
Qaeda? to ?Null? where the individual is not named in the caption, the language model
correctly assigns ?Null? to the face. As table 1 shows, incorporating a language model
improves our face clusters significantly.
the expected values of the set of face-name correspondences (given a face clustering and
language model) and updating the face clusters and language model given the correspondences. Unusually, it is affordable to compute the expected value of all possible face-name
correspondences for a data item since the number of possibilities is small.
To use EM we need a model of how pictures are generated. Generative model:
To generate a data item:
1. Choose N, the number of names, and F, the
number of faces.
2. Generate N name, context pairs.
3. For each of these name, context pairs,
generate a binary variable pictured conditioned on the context alone (from
P (pictured|context)).
4. For each pictured = 1, generate
a face
P
from P (f ace|name) (F n =
pictured).
name, context
pictured
N
face_u
face_n
Fu
Fn
D
5. Generate F u = F ? F n other faces from
P (f ace).
The parameters necessary for the EM process are P (f ace|name) (sec 2.2), the probability
that a face is generated by a given name, P (pictured|context) (sec 2.3), the probability
that a name is pictured given its context, and P (f ace) the probability that a face is generated without a name.
Elizabeth Dole
Angelina Jolie
Elizabeth Dole
Angelina Jolie
Donald Fehr
Donald Fehr
U.S. Open
Abraham Lincoln
Abraham Lincoln
U.S. Open
empty
empty
Anastasia Myskina
Without language model
Anastasia Myskina
With language model
Figure 3: Left: Example clusters using only appearance to cluster. Right: The same
clusters, but using appearance + language to cluster. Some clusters get larger (Elizabeth
Dole, Angelina Jolie) because of the inclusion of more correct faces. Some clusters get
smaller (Anastasia Myskina) because of the exclusion of incorrect faces. All clusters get
more accurate because the language model is breaking ambiguities and giving the clustering a push in the right direction. Some clusters that do not refer to actual people, ?U.S.
Open? completely disappear using the language model. Other clusters like ?Abraham Lincoln? (who is a person, but whose associated pictures most often portray people other than
?Abraham Lincoln?) become empty when using the language model, presumably because
these faces are assigned to the correct names.
2.1
Name Assignment
For each image-caption pair, we calculate the costs of all possible assignments of names to
faces (dependent upon the associated faces and names) and use the best such assignment.
An example of the extracted names, faces and all possible assignments can be seen in figure
1.
The likelihood of picture xi under assignment aj , of names to faces under our generative
model is:
L(xi , aj ) =P (N )P (F )P (n1 , c1 )...P (nn , cn )?
Y
Y
Y
P (pictured? |c? )P (f?(?) |n? )
(1 ? P (pictured? |c? ))
P (f? )
?
?
?
Where P(N) is the probability of generating N names, P(F) is the probability of generating F
faces, and P (ni , ci ) are the probabilities of generating namei and context ci . In assignment
aj , ? indexes into the names that are pictured, ?(?) indexes into the faces assigned to the
pictured names, ? indexes into the names that are not pictured and ? indexes into the faces
without assigned names. The terms P (N )P (F )P (n1 , c1 )...P (nn , cn ) are not dependent
on the assignment so we can ignore them when calculating the probability of an assignment
and focus on the remaining terms.
IN Pete Sampras IN of the U.S. celebrates his victory over Denmark?s OUT Kristian Pless
OUT at the OUT U.S. Open OUT at Flushing Meadows August 30, 2002. Sampras won the
match 6-3 7- 5 6-4. REUTERS/Kevin Lamarque
Germany?s IN Chancellor Gerhard Schroeder IN, left, in discussion with France?s IN President Jacques Chirac IN on the second day of the EU summit at the European Council headquarters in Brussels, Friday Oct. 25, 2002. EU leaders are to close a deal Friday on finalizing entry
talks with 10 candidate countries after a surprise breakthrough agreement on Thursday between
France and Germany regarding farm spending.(AP Photo/European Commission/HO)
?The Right Stuff? cast members IN Pamela Reed IN, (L) poses with fellow cast member IN
Veronica Cartwright IN at the 20th anniversary of the film in Hollywood, June 9, 2003. The
women played wives of astronauts in the film about early United States test pilots and the space
program. The film directed by OUT Philip Kaufman OUT, is celebrating its 20th anniversary
and is being released on DVD. REUTERS/Fred Prouser
Kraft Foods Inc., the largest U.S. food company, on July 1, 2003 said it would take steps, like
capping portion sizes and providing more nutrition information, as it and other companies face
growing concern and even lawsuits due to rising obesity rates. In May of this year, San Francisco attorney OUT Stephen Joseph OUT, shown above, sought to ban Oreo cookies in California ? a suit that was withdrawn less than two weeks later. Photo by Tim Wimborne/Reuters
REUTERS/Tim Wimborne
Figure 4: Our new procedure gives us not only better clustering results, but also a natural
language classifier which can be tested on captions in isolation. Above: a few captions
labeled with IN (pictured) and OUT (not pictured) using our learned language model. Our
language model has learned which contexts have high probability of referring to pictured
individuals and which contexts have low probabilities. We observe an 85% accuracy of
labeling who is portrayed in a picture using only our language model. The top 3 labelings
are all correct. The last incorrectly labels ?Stephen Joseph? as not pictured when in fact
he is the subject of the picture. Some contexts that are often incorrectly labeled are those
where the name appears near the end of the caption (usually a cue that the individual
named is not pictured). Some cues we could add that should improve the accuracy of our
language model are the nearness of words like ?shown?, ?pictured?, or ?photographed?.
The complete data log likelihood is:
?
?
X X
?
(Pij log(L(xi , aj ))?
i?pics
j?ass
Where Pij is an indicator variable telling which correspondence occurred in this data item.
The Pij are missing data whose expectations are computed in the E step.
This gives a straightforward EM procedure:
? E ? update the Pij according to the normalized probability of picture i with assignment j.
? M ? maximize the parameters P (f ace|name) and P (pictured|context) using
soft counts.
2.2
Modeling the appearance of faces ? P (f ace|name)
We model appearance using a mixture model with one mixture element per name in our
lexicon. We need a representation for faces in a feature space where comparisons are
helpful. Our representation is obtained by rectification of the faces followed by kernel
principal components analysis (kPCA) and linear discriminant analysis (LDA) (details in
[5]). We model the distributions P (f ace|name) using gaussians with fixed covariance.
To obtain features we first automatically rectify all faces to a canonical pose. Five support
vector machines are trained as feature detectors (corners of the left and right eyes, corners
Model
Appearance Model, No Lang Model
Appearance Model + Lang Model
EM
56%
72%
MM
67%
77%
Table 1: Above: We randomly selected a set of 1000 faces from our dataset and hand
labeled them with their correct names. Here we show what percentage of those faces are
correctly labeled by each of our methods (EM and maximal correspondence clustering,
MM). For both methods, incorporating a language model improves their respective clusterings greatly. Standard statistical knowledge says that using the expected values should
perform better than simply choosing the maximal assignment at each step (MM). However,
we have found that using the maximal assignment works better than taking an expectation.
One reason this could be true is that EM averages incorrect faces into the appearance
model, making the mean unstable.
of the mouth and the tip of the nose) using features consisting of the geometric blur of [4]
applied to grayscale patches. We then use kPCA ([16]) to reduce the dimensionality of our
data and compute linear discriminants ([3]) on the single name, single face pictures.
Because of the large size of our dataset, we cannot compute the kernel matrix, K for kPCA,
directly. Instead we use an approximation to calculate the eigenvectors of K, the Nystro? m
approximation (cf [17, 9]). The Nystr?om approximation computes two exact subsets of K
and uses these to efficiently approximate the rest of K and its eigenvectors (details in [5]).
2.3
Language Model ? P (pictured|context)
Our language model assigns a probability to each name based on its context within the
caption. These distributions, P (pictured|context), are learned using counts of how often
each context appears describing an assigned name, versus how often that context appears
describing an unassigned name. We have one distribution for each possible context cue,
and assume that context cues are modeled independently (because we lack enough data to
model them jointly).
For context, we use a variety of cues: the part of speech tags of the word immediately
prior to the name and immediately after the name within the caption (modeled jointly), the
location of the name in the caption, and the distances to the nearest ?,?, ?.?, ?(?, ?)?, ?(L)?,
?(R)? and ?(C)?. We tried adding a variety of other language model cues, but found that
they did not increase the assignment accuracy.
The probability of being pictured given multiple context cues (where Ci are the different
independent context cues) can be formed using Bayes rule:
P (pictured|C1 , C2 , ...Cn ) =
P (pictured|C1 )...P (pictured|Cn )
P (pictured)n?1
We compute maximum likelihood estimates of each of P (pictured|Ci ) and P (pictured)
using soft counts.
2.4
Best correspondence and mean correspondence
Given our hidden variable problem of determining correct name-face assignments, from a
statistics point of view EM seems like the most favorable choice. However, many computer
vision problems have observed better results by choosing maximum over expected values.
We have tried both methods and found that using the maximal assignment produced better
results (table 1). One reason this might be true is that for cases where there is a clear best
assignment the max and the average are basically equivalent. For cases where there is no
clear best, EM averages over assignments, producing a mean that has no real meaning since
it is an average of different people?s faces.
Classifier
Baseline
EM Labeling with Language Model
MM Labeling with Language Model
labels correct
67%
76%
84%
IN correct
100%
95%
87%
OUT correct
0%
56%
76%
Table 2: Above: Results of applying our learned language model to a test set of 430
captions (text alone). In our test set, we hand labeled each detected name with IN/OUT
based on whether the referred name was pictured within the corresponding picture. We
then tested how well our language model could predict those labels (?labels correct? refers
to the total percentage of names that were correctly labeled, ?IN correct? the percentage
of pictured names correctly labeled, and ?OUT correct? the percentage of not pictured
names correctly labeled). The baseline figure gives the accuracy of labeling all names
as pictured. Using EM to learn a language model gives an accuracy of 76% while using
a maximum likelihood clustering gives 84%. Again the maximum likelihood clustering
outperforms EM. Names that are most often mislabeled are those that appear near the end
of the caption or in other contexts that usually denote a name being not pictured.
The Maximal Assignment process is nearly the same as the EM process except instead of
calculating the expected value of each assignment only the maximal assignment is nonzero.
The Maximal Assignment procedure:
? M1 ? set the maximal Pij to 1 and all others to 0.
? M2 ? maximize the parameters P (f ace|name) and P (pictured|context) using
counts.
3
Results
We have collected a dataset consisting of approximately half a million news pictures and
captions from Yahoo News over a period of roughly two years.
Faces: Using the face detector of [14], we extract 44,773 large well detected face images.
Since these pictures were taken ?in the wild? rather than under fixed laboratory conditions,
they represent a broad range of individuals, pose, expression, illumination conditions and
time frames. Our face recognition dataset is more varied than any other to date.
Names: We use an open source named entity recognizer ([8]) to detect proper names in
each of the associated captions. This gives us a set of names associated with each picture.
Scale: We obtain 44,773 large and reliable face detector responses. We reject face images
that cannot be rectified satisfactorily, leaving 34,623. Finally, we concentrate on images
within whose captions we detect proper names, leaving 30,281, the final set we cluster on.
3.1
Quantitative Results
Incorporating a natural language model into face clustering produces much better results
than clustering on appearance alone. As can be seen in table 1, using, only appearance
produces an accuracy of 67% while appearance + language gives 77%. For face labeling,
using the maximum likelihood assignment (MM) rather than the average (EM) produces
better results (77% vs 72%).
One neat by-product of our clustering is a natural language classifier. We can evaluate that
classifier on text without associated pictures. In table 2, we show results for labeling names
with pictured and not pictured using our language model. Using the language model we
correctly label 84% of the names while the baseline (labeling everyone as pictured) only
gives 67%. The maximum likelihood assignment also produces a better language model
than EM (76% vs 84%). A few things that our language model learns as indicative of being
pictured are being near the beginning of the caption, being followed by a present tense verb,
and being near ?(L)?, ?(R)?, or ?(C)?.
4
Discussion
We have shown previously ([5]) that a good clustering can be created using names and
faces. In this work, we show that by analyzing language more carefully we can produce
a much better clustering (table 1). Not only do we produce better face clusters, but we
also learn a natural language classifier that can be used to determine who is pictured from
text alone (table 2). We have coupled language and images, using language to learn about
images and images to learn about language.
The next step will be to try to learn a language model for free text on a webpage. One area
we would like to apply this to is improving google image search results. Using a simple
image representation and a modified context model perhaps we could link google images
with the words on the surrounding webpages to improve search results.
References
[1] K. Barnard, D.A. Forsyth, ?Clustering Art,? Computer Vision and Pattern Recognition,
2001
[2] K. Barnard, P. Duygulu, N. de Freitas, D.A. Forsyth, D. Blei, and M.I. Jordan, ?Matching Words and Pictures,? Journal of Machine Learning Research, Vol 3, pp 1107-1135,
2003.
[3] P. Belhumeur, J. Hespanha, D. Kriegman ?Eigenfaces vs. Fisherfaces: Recognition
Using Class Specific Linear Projection? Transactions on Pattern Analysis and Machine
Intelligence, Special issue on face recognition, pp. 711-720, July 1997.
[4] A.C. Berg, J. Malik, ?Geometric Blur for Template Matching,? Computer Vision and
Pattern Recognition,Vol I, pp. 607-614, 2001.
[5] T.L. Berg, A.C. Berg, J. Edwards, M. Maire, R. White, E. Learned-Miller, D.A. Forsyth
?Names and Faces in the News? Computer Vision and Pattern Recognition, 2004.
[6] V. Blanz, T. Vetter, ?Face Recognition Based on Fitting a 3D Morphable Model,?
Transactions on Pattern Analysis and Machine Intelligence Vol. 25 no.9, 2003.
[7] C. Carson, S. Belongie, H. Greenspan, J. Malik, ?Blobworld ? Image segmentation using expectationmaximization and its application to image querying,? IEEE Transactions
on Pattern Analysis and Machine Intelligence, 24(8), pp. 1026?1038, 2002.
[8] H. Cunningham, D. Maynard, K. Bontcheva, V. Tablan, ?GATE: A Framework and
Graphical Development Environment for Robust NLP Tools and Applications,? 40th
Anniversary Meeting of the Association for Computational Linguistics?, Philadelphia,
July 2002.
[9] C. Fowlkes, S. Belongie, F. Chung, J. Malik, ?Spectral Grouping Using The Nystr o? m
Method,? TPAMI, Vol. 26, No. 2, February 2004.
[10] R. Gross, J. Shi and J. Cohn, ?Quo Vadis Face Recognition?,? Third Workshop on
Empirical Evaluation Methods in Computer Vision, December, 2001.
[11] R. Gross, I. Matthews, and S. Baker, ?Appearance-Based Face Recognition and LightFields,? Transactions on Pattern Analysis and Machine Intelligence, 2004.
[12] V. Lavrenko, R. Manmatha., J. Jeon, ?A Model for Learning the Semantics of Pictures,? Neural Information Processing Systems, 2003
[13] J. Li and J. Z. Wang, ?Automatic Linguistic Indexing of Pictures by a Statistical Modeling Approach,? Transactions on Pattern Analysis and Machine Intelligence, vol. 25,
no. 9, pp. 1075-1088, 2003
[14] K. Mikolajczyk ?Face detector,? Ph.D report, INRIA Rhone-Alpes
[15] J. Scheeres,
?Airport face scanner failed?,
Wired News,
2002.
http://www.wired.com/news/privacy/0,1848,52563,00.html.
[16] B. Scholkopf, A. Smola, K.-R. Muller ?Nonlinear Component Analysis as a Kernel
Eigenvalue Problem? Neural Computation, Vol. 10, pp. 1299-1319, 1998.
[17] C. Williams, M. Seeger ?Using the Nystro? m Method to Speed up Kernel Machines,?
Advances in Neural Information Processing Systems, Vol 13, pp. 682-688, 2001.
| 2708 |@word middle:2 rising:1 seems:1 open:6 tried:2 covariance:1 nystr:2 manmatha:1 quo:1 united:2 pless:1 outperforms:1 dole:3 freitas:1 com:1 lang:2 john:2 fn:1 blur:2 treating:2 concert:1 update:1 v:3 alone:5 cue:9 half:2 generative:2 item:8 selected:1 indicative:1 intelligence:5 beginning:3 blei:1 nearness:1 provides:1 iterates:2 cheney:1 lexicon:1 location:1 rhone:1 lavrenko:1 five:1 c2:1 become:1 scholkopf:1 incorrect:5 combine:1 wild:1 fitting:1 privacy:1 expected:5 roughly:1 growing:1 detects:1 finalizing:1 automatically:2 food:2 actual:3 company:2 baker:1 null:21 what:1 kaufman:1 string:1 corporation:1 berkeley:3 fellow:1 quantitative:1 nation:1 unusually:2 stuff:1 classifier:6 astronaut:1 appear:3 organize:1 producing:1 before:13 analyzing:1 approximately:2 ap:1 might:1 inria:1 studied:1 suggests:1 co:1 range:1 qaeda:2 directed:1 satisfactorily:1 practice:1 nystro:2 procedure:6 maire:1 area:1 empirical:1 significantly:2 reject:1 matching:3 projection:1 word:8 vetter:1 donald:2 refers:1 suggest:2 get:3 cannot:3 close:1 context:34 applying:1 www:1 equivalent:1 center:1 missing:1 shi:1 straightforward:1 williams:1 independently:1 alpes:1 assigns:3 immediately:2 m2:1 rule:1 jackson:3 his:1 variation:1 president:4 gerhard:1 caption:28 exact:1 us:1 agreement:1 element:1 recognition:14 updating:2 summit:3 labeled:11 observed:1 wang:1 calculate:2 region:1 news:8 olympics:3 eu:2 gross:2 environment:1 kriegman:1 trained:1 upon:1 division:1 f2:9 kraft:1 completely:1 mislabeled:1 joint:1 talk:1 surrounding:2 detected:4 labeling:8 kevin:1 choosing:2 quite:2 ace:8 larger:1 whose:3 film:3 say:1 blanz:1 statistic:1 farm:1 jointly:2 final:1 tpami:1 questioning:1 eigenvalue:1 maximal:8 product:1 combining:1 date:1 lincoln:4 webpage:2 cluster:14 empty:3 produce:10 generating:3 wired:2 tim:2 andrew:1 pose:3 nearest:1 expectationmaximization:1 edward:2 c:1 marcel:3 direction:1 concentrate:1 annotated:1 correct:12 assign:1 f1:9 mm:5 scanner:1 withdrawn:1 presumably:1 week:1 predict:1 matthew:1 sought:1 early:1 released:1 recognizer:6 favorable:2 failing:1 bag:1 bond:1 label:5 council:1 him:1 largest:1 vice:1 hollywood:1 tool:1 weighted:1 brought:1 modified:1 rather:2 unassigned:1 greenspan:1 linguistic:1 focus:1 june:1 maria:2 likelihood:9 indicates:1 greatly:1 seeger:1 baseline:3 detect:2 helpful:1 rear:1 dependent:2 nn:2 unlikely:2 cunningham:1 hidden:3 captioned:1 her:1 france:2 labelings:1 germany:2 semantics:1 headquarters:1 issue:1 html:1 yahoo:1 development:1 art:1 breakthrough:1 fairly:1 special:1 airport:1 once:1 ioc:1 broad:1 nearly:1 jon:1 thinking:1 others:1 report:1 few:3 fehr:2 winter:3 randomly:1 museum:1 individual:6 floating:1 consisting:3 jeon:1 n1:11 suit:1 detection:1 possibility:2 evaluation:1 punctuation:1 mixture:2 accurate:2 fu:1 necessary:1 respective:1 calif:1 cooky:1 instance:1 soft:2 modeling:2 stewart:1 assignment:26 maximization:1 kpca:4 cost:1 entry:1 subset:1 commission:1 referring:1 person:1 michael:3 tip:1 again:1 ambiguity:1 choose:2 woman:1 corner:2 american:1 chung:1 friday:2 return:1 li:1 potential:2 de:1 attorney:1 sec:2 inc:1 forsyth:4 depends:1 later:1 view:1 try:1 portion:1 bayes:1 annotation:1 om:1 formed:1 ni:1 accuracy:6 who:5 efficiently:1 miller:1 identify:2 raw:1 accurately:1 produced:1 basically:1 rectified:1 avram:3 detector:6 against:2 pp:7 tamara:1 james:2 associated:10 gain:1 pilot:1 dataset:7 knowledge:1 improves:2 dimensionality:1 segmentation:1 carefully:2 appears:5 ta:1 supervised:2 day:1 response:1 smola:1 hand:2 cohn:1 nonlinear:1 lack:1 google:2 petr:1 maynard:1 mode:1 lda:2 aj:4 perhaps:1 name:80 tense:2 normalized:1 true:2 assigned:9 read:1 laboratory:2 nonzero:1 deal:1 white:1 round:1 during:2 won:2 carson:1 complete:1 julia:4 image:19 spending:1 meaning:1 common:2 superior:1 discriminants:1 million:3 linking:1 he:1 occurred:1 ukrainian:1 m1:1 association:1 significant:1 refer:3 automatic:1 inclusion:1 language:49 rectify:1 operating:1 meadow:1 morphable:1 pete:1 add:1 exclusion:1 confounding:1 occasionally:3 browse:1 binary:1 meeting:1 muller:1 seen:2 george:1 belhumeur:1 determine:1 maximize:2 period:1 july:4 semi:1 heidi:1 multiple:3 stephen:2 match:1 victory:2 vision:5 expectation:4 affordable:1 kernel:4 represent:1 c1:4 country:1 source:1 leaving:2 rest:1 subject:1 thing:1 member:2 december:1 smile:1 prague:2 jordan:1 near:5 enough:1 bid:2 variety:2 isolation:3 reduce:1 regarding:1 cn:4 court:1 whether:1 expression:1 speech:1 afford:1 prefers:1 santa:1 eigenvectors:2 clear:2 ph:1 lawsuit:2 generate:5 http:1 percentage:4 canonical:1 jacques:1 correctly:7 per:1 vol:7 officer:1 jaety:1 year:2 wife:1 powerful:1 named:8 patch:1 ric:1 followed:6 played:1 correspondence:12 schroeder:1 occur:1 dvd:1 tag:2 speed:1 duygulu:1 photographed:1 according:2 brussels:1 watt:1 ball:1 smaller:1 em:19 elizabeth:3 joseph:2 making:2 indexing:1 gathering:1 taken:1 rectification:1 previously:1 jennifer:4 describing:2 count:4 nose:1 end:2 photo:2 available:2 gaussians:1 opponent:1 apply:1 observe:1 spectral:1 occurrence:1 fowlkes:1 ho:1 gate:1 top:1 clustering:17 ceo:2 remaining:1 cf:1 graphical:1 nlp:1 linguistics:1 calculating:2 giving:1 disappear:1 february:1 malik:3 cartwright:1 responds:1 anastasia:3 said:1 exhibit:1 distance:1 link:2 entity:6 philip:1 blobworld:1 collected:2 discriminant:1 unstable:1 reason:2 nutrition:1 denmark:1 index:4 reed:1 modeled:2 providing:1 dick:1 difficult:1 hespanha:1 proper:2 fisherfaces:1 perform:1 observation:1 datasets:1 incorrectly:2 looking:1 frame:1 varied:1 verb:3 august:1 ivory:1 pic:1 david:1 pair:5 paris:1 cast:2 california:1 learned:6 usually:2 pattern:8 program:1 max:1 reliable:1 everyone:1 mouth:1 natural:7 indicator:1 pictured:42 improve:3 millennium:1 eye:1 picture:21 identifies:1 created:1 portray:1 obesity:1 extract:1 columbia:1 coupled:1 philadelphia:1 text:4 review:1 geometric:2 prior:1 vancouver:2 chancellor:1 determining:1 fully:1 querying:1 versus:1 remarkable:1 pij:5 preposition:2 anniversary:3 ban:1 last:1 free:1 neat:1 telling:1 eigenfaces:1 face:75 taking:1 template:1 world:1 fred:1 computes:1 mikolajczyk:1 collection:3 san:2 transaction:5 nov:1 approximate:1 ignore:1 belongie:2 naomi:1 francisco:1 xi:3 leader:1 grayscale:1 search:2 chief:1 table:8 learn:6 reasonably:1 robust:1 ca:1 obtaining:1 improving:1 as:1 european:2 domain:1 did:1 abraham:4 motivation:1 reuters:5 n2:9 fair:1 complementary:1 referred:1 surveyed:1 candidate:1 crude:1 house:1 breaking:1 portrayed:1 late:1 third:1 capping:1 learns:1 british:1 specific:1 veronica:1 concern:1 grouping:1 incorporating:5 workshop:1 adding:1 ci:4 pierce:1 illumination:1 conditioned:1 push:1 easier:1 surprise:1 pamela:1 depicted:3 photograph:1 simply:1 appearance:16 likely:1 failed:1 furlong:2 kristian:1 extracted:1 oct:1 barnard:2 man:1 martha:1 typical:1 except:1 principal:1 total:1 vote:2 thursday:2 berg:5 mark:1 people:5 support:1 alexander:1 evaluate:2 tested:2 phenomenon:1 |
1,882 | 2,709 | Semigroup Kernels on Finite Sets
Marco Cuturi
Computational Biology Group
Ecole des Mines de Paris
35 rue Saint Honor?e
77300 Fontainebleau
[email protected]
Jean-Philippe Vert
Computational Biology Group
Ecole des Mines de Paris
35 rue Saint Honor?e
77300 Fontainebleau
[email protected]
Abstract
Complex objects can often be conveniently represented by finite sets of
simpler components, such as images by sets of patches or texts by bags
of words. We study the class of positive definite (p.d.) kernels for two
such objects that can be expressed as a function of the merger of their
respective sets of components. We prove a general integral representation of such kernels and present two particular examples. One of them
leads to a kernel for sets of points living in a space endowed itself with a
positive definite kernel. We provide experimental results on a benchmark
experiment of handwritten digits image classification which illustrate the
validity of the approach.
1
Introduction
Suppose we are to deal with complex (e.g non-vectorial) objects from a set Z on which
we wish to apply existing kernel methods [1] to perform tasks such as classification or
regression. Assume furthermore that the latter objects can be meaningfully described by
small components contained in a set X . Namely, we suppose that we can define an a
priori mapping ? which maps any z ? Z into a finite unordered list of elements of X ,
? (z) = [x1 , x2 , ..., xn ], through a sampling process which may be exhaustive, heuristic or
random both in the quantity of sampled components n and in the way those components
are extracted. Comparing two such complex objects through the direct comparison of their
respective lists of components has attracted much attention recently, namely through the
definition of p.d. kernel on such ? -lists. Most recent approaches to compare two ? -lists
involve the estimation of two distributions pz and pz? on X within a parametric class of
models that fit (e.g. in maximum likelihood (ML) sense) respectively ? (z) and ? (z ? ) seen
as a samples from laws on X , where each resulting law could be identified with z and z ?
respectively. Such a kernel is then defined between pz and pz? , as seen for example in [2]
with the Information Diffusion kernel, in [3] with the family of Mutual Information Kernels
or in [4] with the use of the Battacharyya affinity between pz and pz? . An alternative
and non-parametric approach to ? -lists comparison that studies the subspaces generated by
points of ? (z) and ? (z ? ) in a feature space was also proposed in [5], recalling elements
presented in Kernel-Canonical Correlation Analysis [6].
We explore in this contribution a different direction to kernel design for lists by studying the
class of kernels whose value computed on two lists is only defined through its value on their
concatenation. This approach was already used in [7], where a particular kernel for strings
that only compares two strings through their concatenation is presented. In this paper, the
approach is extended to a more general and abstract setting of ? -lists, but the motivation
remains the same as in [7]: if two ? -lists are similar, e.g. in terms of the distribution of
the components they describe, then their concatenation will be more ?concentrated? than if
they are very different, in which case it might look more like a reunion of two disjoint sets
of points. As a result, one can expect to get a relevant measure of similarity, and hence a
kernel, by studying properties of the concatenation of two lists such as its concentration.
After an example of a valid kernel for lists seen as measures on the space of components
(Section 2), we provide a complete characterization for this class of kernels (Section 3)
by casting them in the context of semigroup kernels. This leads to the definition of a
second kernel based on exponential densities on X , which boils down after a numerical
approximation to the computation of the entropy of the maximum likelihood density (taken
in the considered exponential family) of the points numbered by lists of components. This
kernel is extended in Section 4 to points taken in a reproducing kernel Hilbert space defined
by a kernel ? on X , and is then tested on a problem of image classification, where images
are seen as bags of pixels and a non-linear kernel between pixels is used (Section 5).
2
The entropy kernel
As a warm-up, let us assume that the set X is measurable, e.g. X = Rd , and that to
any point x ? X we can associate a probability measure on X with density ?x with
respect to a common measure (e.g. the Borel uniform measure), with finite entropy
R
def
h(?) = ? X ? ln ?. Consider for example a Gaussian distribution with mean x and
n
fixed variance. A natural way
Pnto represent an unordered list ? (z) = [x1 , x2 , ..., xn ] ? X
is by the density ?? = 1/n i=1 ?xi . In that case, a p.d. kernel k between unordered lists
? and ? ? that only depends on their concatenation ? (z) ? ? (z ? ) is equivalent to a p.d. kernel
between densities ? and ?? that only depends on ?+?? . Hence we are looking for a p.d. kernel on the set P of probability densities of finite entropy of the form ?(?, ?? ) = ?(? + ?? ).
An example of such a kernel is provided in the following proposition. Recall that a negative
g : X 2 ? R that satisfies
Pn definite (n.d.) kernel on a set X is a symmetric function
n
c
c
g(x
,
x
)
?
0
for
any
n
?
N,
(x
,
.
.
.
,
x
)
?
X
,
and
(c1 . . . , cn ) ? Rn with
i j
i
j
1
n
Pi,j=1
n
i=1 ci = 0. A useful link between p.d. and n.d. kernels is that g is n.d. if and only if
exp(?tg) is p.d. for all t > 0 [8, Theorem 3.2.2.].
?
Proposition 1. The function g : ?, ?? 7? h( ?+?
2 ) is negative definite on P, making
def
?+??
kh (?, ?? ) = e?th( 2
between two measures.
)
a p.d. kernel on P for any t > 0. We call kh the entropy kernel
The entropy kernel is already a satisfactory answer to our initial motivation to look at
merger of points. Observe that if ?x is a probability density around x, then ?? can often
be thought of as an estimate of the distribution of the points in ? , and (?? + ?? ? )/2 is an
estimate of the distribution of the points enumerated in ? ? ? ? . If the latter estimate has a
small entropy we can guess that the points in ? and ? ? are likely to have similar distributions
which is exactly the similarity that is quantified by the entropy kernel.
Proof of Proposition 1. It is known that the real-valued function r : y 7? ?y ln y is n.d.
on R+ as a semigroup endowed with addition [8, Example 6.5.16]. As a consequence the
function f 7? r ? f is n.d. on P as a pointwise application of r, and so is its summation on
X . For any real-valued n.d. kernel k and any real-valued function g, we have trivially that
?
(y, y ? ) 7? k(y, y ? ) + g(y) + g(y ? ) remains negative definite, hence h( f +f
2 ) is n.d. through
?
1
ln 2
?
?
h( f +f
2 ) = 2 h(f + f ) + 2 (|f | + |f |), yielding positive definiteness of kh . ?
3
Semigroups and integral representations of p.d. kernels on finite
Radon measures
In order to generalize the example presented in the previous section, let us briefly recall
the concept of p.d. kernels on semigroups [8]. A nonempty set S is called an Abelian
(autoinvolutive) semigroup if it is equipped with an Abelian associative composition ?
admitting a neutral element in S. A function ? : S 7? R is called a positive definite
(resp. negative definite) function on the semigroup (S, ?) if (s, t) 7? ?(s ? t) is a p.d. (resp.
n. d.) kernel on S ? S.
The entropy kernel defined in Proposition 1 is therefore a p.d. kernel on the semigroup
of measures with finite entropy endowed with usual addition. This can be generalized by
assuming that X is a Hausdorff space, which suffices to consider the set of finite Radon
b
b
measures M+
(X ) [8]. For ? ? M+
(X ), we Rnote |?| = ?(X ) < +?. For a Borel
X
measurable function f ? R , we note ?[f ] = X f d?. Endowed with the usual Abelian
b
addition between measures, (M+
(X ), +) is an Abelian semigroup. The reason to consider
this semigroup is that there is a natural semigroup homomorphism between
Pn finite lists of
b
points and elements of M+
(X ) given by ? = [x1 , ..., xn ] 7? ?? =
i=1 ?xi , where
b
?x ? M+
(X ) is an arbitrary finite measure associated with each x ? X . We discussed in
section 2 the case where ?x has a density, but more general measures are allowed, such as
?x = ?x , the Dirac measure. Observe that when we talk about lists, it should be understood
that some objects might appear with some multiplicity which should be taken
account
Pinto
n
(specially when X is finite), making us consider weighted measures ? = i=1 ci ?xi in
the general case. We now state the main result of this section which characterizes bounded
b
p.d. functions on the semigroup M+
(X ),
b
Theorem 1. A bounded real-valued function ? on M+
(X ) such that ?(0) = 1 is p.d. if
and only if it has an integral representation:
Z
?(?) =
e??[f ] d?(f ),
C + (X )
where ? is a uniquely determined positive radon measure on C + (X ), the space of nonnegative-valued continuous functions of RX endowed with the topology of pointwise convergence.
b
Proof. (sketch) Endowed with the topology of weak convergence, M+
(X ) is a Hausdorff
space [8, Proposition 2.3.2]. The general result of integral representation of bounded p.d.
function [8, Theorem 4.2.8] therefore applies. It can be shown that bounded semicharacters
b
on M+
(X ) are exactly the functions of the form ? 7? exp(??[f ]) where f ? C + (X ) by
using the characterization of semicharacters on (R+ , +) [8, Theorem 6.5.8] and the fact
b
that atomic measures is a dense subset of M+
(X ) [8, Theorem 2.3.5].
As a constructive application to this general representation theorem, let us consider the
case ?x = ?x and consider, as a subspace of C + (X ), the linear span of N non-constant,
continuous, real-valued and linearly independent functions f1 , ..., fN on X . As we will see
below, this is equivalent to considering a set of densities defined by an exponential model,
PN
namely of the form p? (x) = exp( j=1 ?j fj (x)??(?)) where ? = (? j )j=1..N ? ? ? RN
is variable and ? is a real-valued function defined on ? to ensure normalization of the
densities p? . Considering a prior ? on the parameter space ? is equivalent to defining
a Radon measure taking positive values on the subset of C + (X ) spanned by functions
f1 , ..., fN . We now have ( see [9] for a geometric point of view) that:
Theorem 2. ??? ? ? being the ML parameter associated with ? and noting p? = p??? ,
Z
?|?|h(p? )
?? (?) = e
e?|?|d(p? ||p? ) ?(d?),
?
is a p.d. kernel on the semigroup of measures, where d(p||q) =
Kullback-Leibler divergence between p and q.
R
supp(q)
p ln pq is the
Although an exact calculation of the latter equation is feasible in certain cases (see [10, 7]),
an approximation can be computed using Laplace?s approximation. If for example the prior
on the densities is taken to be Jeffrey?s prior [9, p.44] then the following approximation
holds:
? ? N2
2?
.
(1)
?(?) ? ?(?)
?
:= e?|?|h(p? )
|?|
|?|??
2
The ML estimator being unaffected by the total weight |?|, we have ?(2?)
?
= ?(?)
? 2 ( |?|
4? )
which we use to renormalize our kernel on its diagonal:
? p
! N2
?(|?+?? |)h(p?+?? )
?|
2
|?||?
e
k(?, ?? ) = ?|?|h(p )?|?? |h(p ? )
?
?
|?| + |?? |
e
N
Two problems call now for a proper renormalization: First, if |?? | ? |?| (which would
be the case if ? describes far more elements than ? ? ), the entropy h(p?+?? ) will not take
into account the elements enumerated in ?? . Second, the value taken by our p.d function ??
decreases exponentially with |?| as can be seen in equation (1). This inconvenient scaling
behavior leads in practice to bad SVM classification results due to diagonal dominance of
the Gram matrices produced by such kernels (see [11] for instance). Recall however that
the Laplace approximation can be accurate only when |?| ? 0. To take into account this
tradeoff on the ideal range of |?|, we rewrite the previous expression using a common width
parameter ? after having applied a renormalization on ? and ?? :
?2?
?
?
k? (?, ? ) = k( ?, ? ?? ) = e
|?| |? |
?
?
h(p??? )?
h(p? )+h(p ? )
?
2
?
,
(2)
?
?
+ |??? | . ? should hence be big enough in practical applications to ensure the
where ??? = |?|
consistency of Laplace?s approximation and thus positive definiteness, while small enough
to avoid diagonal dominance. We will nowP
always suppose that our atomic measures are
n
normalized, meaning that their total weight i=1 ci always sums up to 1.
Let us now review a practical case when X is Rk , and that some kind of gaussianity among
points makes sense. We can use k-dimensional normal distributions pm,? ? N (m, ?)
(where ? is a k ? k p.d. matrix)
Pn to define our densities.
Pn The ML parameters?of a measure
? are in that case : ?
? = i=1 ci xi and ?? = i=1 ci (xi ? ?
?)(xi ? ?
?) . Supposing
that the span of the n vectors xi covers Rk yields non-degenerated covariance matrices.
This ensures the existence of the entropy of the ML estimates through the formula [12]:
h(pm,? ) = 12 ln ((2?e)n |?|). The value of the normalized kernel in (2) is then:
?p
!2?
|?? ||??? |
?
k? (?, ? ) =
.
|???? |
This framework is however limited to vectorial data for which the use of Gaussian laws
makes any sense. An approach designed to bypass this double restriction is presented in
the next section, taking advantage of a prior knowledge on the components space through
the use of a kernel ?.
4
A kernel defined through regularized covariance operators
Endowing X (now also considered 2-separable) with a p.d. kernel ? bounded on the diagonal, we make use in this section of its corresponding reproducing kernel Hilbert space
(RKHS, see [13] for a complete survey). This RKHS is denoted by ?, and its feature map
by ? : x 7? ?(x, ?). ? is infinite dimensional in the general case, preventing any systematical use of exponential densities on that feature space. We bypass this issue through
a generalization of the previous section by still assuming some ?gaussianity? among the
elements numbered by atomic measures ?, ?? and ??? which, once mapped in the feature
space, are now functions. More precisely, our aim when dealing with Euclidean spaces
was to estimate finite dimensional covariance matrices ?? , ??? , ???? and compare them in
terms of their spectrum or more precisely through their determinant. In this section we
use such finite samples to estimate, diagonalize and regularize three covariance operators
S? , S?? , S??? associated with each measure on ?, and compare them by measuring their respective dispersion in a similar way. We note for ? ? ? its dual ? ? (namely the linear form
? ? R s.t. ? 7? ? ? ? = h?, ?i? ) and ||?||2 = ? ? ?. Let (ei )i?N be a complete orthonormal
base of ? (i.e. such that span(ei )i?N = ? and e?i ej = ?ij ). Given a family of positive real
?
numbers (ti )i?N
P, we note?St,e the bilinear symmetric operator which maps ?, ? 7? ? St,e ?
where St,e = i?N ti ei ei .
def
For an atomic measure ? and notingP
??i = (?i ? ?[?]) its n centered points in ?, the
n
empirical covariance operator S? = i=1 ci ?i ?i? on ? can be described through such a
diagonal representation by finding its principal eigenfunctions, namely orthogonal functions in ? which maximize the expected (w.r.t to ?) variance of the normalized dot-product
def v ? ?
hv (?) = ||v||
here defined for any v of ?. Such functions can be obtained through the
following recursive maximizations:
n
1 X ??
ci vj ?i .
vj =
argmax
var? (hv (?)) =
argmax
2
v??,v?{v1 ,...,vj?1 }
v??,v?{v1 ,...,vj?1 } ||vj || i=1
As in the framework of Kernel PCA [1] (from which this calculus only differs by considering weighted points in the feature space) we have by the representer theorem [1] that all the
solutions of these successive maximizations lie in span({??i }i=1..n ). Thus for each vj there
Pd
? ? ?j where K
?? =
exists a vector ?j of Rn such that vj = i=1 ?j,i ??i with ||vj ||2 = ?j? K
(In ? 1n,n ?c )K? (In ? ?c 1n,n ) is the centered Gram matrix K? = [?(xi , xj )]1?i,j?n of
the points taken in the support of ?, with 1n,n being the n ? n matrix composed of ones
and ?c the n ? n diagonal matrix of ci coefficients. Our latter formulation is however
? ? . We
ill-defined, since any ?j is determined up to the addition of any element of ker K
def
? ? ? Rn to consider functions of positive
thus restrict our parameters ? to lie in E = ker K
?
squared norm, having now:
?j =
? ? ?c K
? ?? ?
?
?? K
= var? (hvj (?))
?
?
? K? ?
? k =0
??E:?k<j,?? K?
argmax
? ? ?c K
? ? and K
? ? being symmetric positive definite on E (one can
Both endomorphism K
?
?
? ? ), the right-hand argument of the previous equaeasily prove that ker K? = ker K? ?c K
? ? ?c K
? ? by K
? ? , can be maximized through a
tion, known as the Rayleigh quotient of K
Hermitian generalized eigenvalue decomposition. This computation yields a basis ?j of E
? ? ?i = 0 for i < j ? dim(E), and with corresponding positive eigenvalues
such that ?j? K
Pd
in decreasing order ?1 , ..., ?dim(E) . Through vj = i=1 ?j,i ??i and writing r = dim(E),
this also yields an orthogonal basis (vj )i?r of span{(??i )i?n }, which can be completed
to span ? through a Gram-Schmidt orthonormalization process using the original basis
(ei )i?N . The orthonormal base corresponding to S? is thus (vi )i?N , where the r first vectors are the original eigenvectors obtained through the previous maximization. Such a
diagonal representation of S? takes the form S? = S?,v where ? = (?1 , ..., ?r , 0, ...). This
bilinear form is however degenerated on ? and facing the same problem encountered in
[4, 6] we also propose to solve this issue through a regularization by adding a component
? on every vector of the base, i.e. defining ?? = (?1 + ?, ..., ?r + ?, ?, ...) with ? > 0, to
propose a regularization of S? as:
S?? ,v =
r
X
i=1
(?i + ?)vi vi? +
X
? vi vi? .
i>r
The entropy of a covariance operator St,e not being defined, we bypass this issue by considering the entropy of its marginal distribution on its first d eigenfunctions, namely introducPd
ing the quantity |St,e |d = d2 ln(2?e) + 12 i=1 ln ti . Let us sum up ideas now and consider
?
three normalized measures ?, ?? and ??? = ?+?
2 , which yield three different orthonormal
bases vi , vi? and vi?? of ? and three different families of weights ?? = (?i?r +?, ?, ...), ??? =
(??i?r? + ?, ?, ...) and ???? = (???i?r?? + ?, ?, ...). Though working on different bases, those
respective d first directions allow us to express an approached form of kernel (2) limited to
different subspaces of ? of arbitrary size d ? r?? ? max(r, r? ):
?
?
!!
|S?? ,v |d + |S??? ,v? |d
?
kd,? (?, ? ) = exp ?2? |S???? ,v?? |d ?
2
? qQ
?
2?
(3)
??i
r
?i Qr ?
1
+
1
+
i=1
i=1
?
?
? ,
=?
Qr??
???
i
i=1 1 + ?
The latter expression is independent of d, while letting d go to infinity lets every base on
which are computed our entropies span the entire space ?. Though the latter hint does not
establish a valid theoretical proof of the positive definiteness of this kernel, we use this final
formula for the following classification experiments.
5
Experiments
Following the previous work of [4], we have conducted experiments on an extraction of
500 images (28 ? 28 pixels) taken in the MNIST database of handwritten digits, with 50
images for each digit. To each image z we randomly associate a set ? (z) of 25 to 30
pixels among black points (intensity superior to 191 on a 0 to 255 scale ) in the image,
where X is {1, .., 28} ? {1, .., 28} in this case. In all our experiments we set ? to be 21
which always yielded positive definite Gram matrices in practice. To define our RKHS
? we used both the linear kernel, ?a ((x1 , y1 ), (x2 , y2 )) = (x1 x2 + y1 y2 )/272 and the
2
2
1 ?y2 )
Gaussian kernel of width ?, namely ?b ((x1 , y1 ), (x2 , y2 )) = e? (x1 ?x227) 2+(y
. The
?2? 2
linear case boils down to the simple application presented in the end of section 3 where we
fit Gaussian bivariate-laws on our three measures and define similarity through variance
analysis. The resulting diagonal variances (?1,1 , ?2,2 ),(??1,1 , ??2,2 ) and (???1,1 , ???2,2 ) measure
of our data for each of the three measures, yielding a kernel value of
? the dispersion
?1,1 ?2,2 ??1,1 ??2,2
equal to 0.382 in the case shown in figure 1. The linear kernel man??
???
1,1 ?2,2
ages a good discrimination between clearly defined digits such as 1 and 0 but fails at
doing so when considering numbers whose pixels? distribution cannot be properly characterized by ellipsoid-like shapes. Using instead the Gaussian kernel brings forward a
non-linear perspective to the previous approach since it maps now all pixels into Gaussian bells, providing thus a much richer function class for ?. In this case two parameters
(a)
?1,1 = 0.0552
?2,2 = 0.0013
(b)
?1 = 0.276
??1,1 = 0.0441
??2,2 = 0.0237
???1,1 = 0.0497
???2,2 = 0.0139
??1 = 0.168
???1 = 0.184
0
Figure 1: First Eigenfunction of three empirical measures ?1 , ?0 and ?1 +?
using the linear
2
(a) and the Gaussian (b, with ? = 0.01, ? = 0.1) kernel. Below each image are the corresponding eigenvalues which correspond to the variance captured by each eigenfunction,
the second eigenvalue being also displayed in the linear case (a).
require explicit tuning: ? (the width of ?) controls the range of the typical eigenvalues
found in the spectrum of our regularized operators whereas ? acts as a scaling parameter for the latter values as can be seen in equation (3). An efficient choice can thus only
be defined on pairs of parameter, which made us use two ranges of values for ? and ?
based on preliminary attempts: ? ? 10?2 ? {0.1, 0.3, 0.5, 0.8, 1, 1.5, 2, 3, 5, 8, 10, 20} and
? ? 10?1 ?{0.5, 1, 1.2, 1.5, 1.8, 2, 2.5, 3}. For each kernel computed on the base of a (?, ?)
couple, we used a balanced training fold of our dataset to train 10 binary SVM classifiers,
namely one for each digit versus all other 9 digits. The class of the remaining images of the
test fold was then predicted to be the one with highest SVM score among the the 10 previously trained binary SVMs. Splitting our data into test and training sets was led through
a 3-fold cross validation (roughly 332 training images and 168 for testing), averaging the
test error on 5 random fold splits of the original data. Those results were obtained using
the spider toolbox1 and graphically displayed in figure (2). Note that the best testing errors
were reached using a ? value of 0.12 with an ? parameter within 0.008 and 0.02, this error
being roughly 19.5% with a standard deviation inferior to 1% in all the region corresponding to an error lower than 22%. To illustrate the sensibility of our method to the number of
sampled points in ? we show in the same figure the decrease of this error when the number
of sampled points ranges from 10 to 30 with independently chosen random points for each
computation. As in [4], we also compared our results to the standard RBF kernel on images
seen as vectors of {0, 1}27?27 , using a fixed number of 30 sampled points and the formula
||z?z ? ||
k(z, z ? ) = e? 30?2?2 . We obtained similar results with an optimal error rate of roughly
44.5% for ? ? {0.12, 0.15, 0.18}. Our results didn?t improve by choosing different soft
margin C parameters, which we hence just set to be C = ? as is chosen by default by the
spider toolbox.
1
see http://www.kyb.tuebingen.mpg.de/bs/people/spider/
50%
2
10 ?
0.1
0.3
0.5
0.8
1
1.5
2
3
5
8
10
20
45%
0.05
e < 22 %
e < 19.5 %
0.12
0.15
?
e < 22 %
0.18
40%
Averaged error rate
0.1
35%
30%
25%
0.2
20%
0.25
15%
10
0.3
(a)
15
20
# points
25
30
(b)
Figure 2: (a) Average test error (displayed as a grey level) of different SVM handwritten
character recognition experiments using 500 images from the MNIST database (each seen
as a set of 25 to 30 randomly selected black pixels), carried out with 3-fold (2 for training, 1
for test) cross validations with 5 repeats, where parameters ? (regularization) and ? (width
of the Gaussian kernel) have been tuned to different values. (b) Curve of the same error
(with ? = 0.01, ? = 0.12 fixed) depending now on the size of the sets of randomly selected
black pixels for each image, this size varying between 10 and 30.
Acknowledgments
The authors would like to thank Francis Bach, Kenji Fukumizu and J?er?emie Jakubowicz
for fruitful discussions and Xavier Dupr?e for his help on the MNIST database.
References
[1] B. Sch?olkopf and A.J. Smola. Learning with Kernels: Support Vector Machines, Regularization,
Optimization, and Beyond. MIT Press, Cambridge, MA, 2002.
[2] J. Lafferty and G. Lebanon. Information diffusion kernels. In Advances in Neural Information
Processing Systems 14, Cambridge, MA, 2002. MIT Press.
[3] M. Seeger. Covariance kernels from bayesian generative models. In Advances in Neural Information Processing Systems 14, pages 905?912, Cambridge, MA, 2002. MIT Press.
[4] R. Kondor and T. Jebara. A kernel between sets of vectors. In Machine Learning, Proceedings
of the Twentieth International Conference (ICML 2003), pages 361?368. AAAI Press, 2003.
[5] L. Wolf and A. Shashua. Learning over sets using kernel principal angles. Journal of Machine
Learning Research, 4:913?931, 2003.
[6] F. Bach and M. Jordan. Kernel independent component analysis. Journal of Machine Learning
Research, 3:1?48, 2002.
[7] M. Cuturi and J.-P. Vert. A mutual information kernel for sequences. In IEEE International
Joint Conference on Neural Networks, 2004.
[8] C. Berg, J.P.R. Christensen, and P. Ressel. Harmonic Analysis on Semigroups. Springer, 1984.
[9] S. Amari and H. Nagaoka. Methods of information geometry. AMS vol. 191, 2001.
[10] F. M. J. Willems, Y. M. Shtarkov, and Tj. J. Tjalkens. The context-tree weighting method: basic
properties. IEEE Transancations on Information Theory, pages 653?664, 1995.
[11] J.-P. Vert, H. Saigo, and T. Akutsu. Local alignment kernels for protein sequences. In
B. Schoelkopf, K. Tsuda, and J.-P. Vert, editors, Kernel Methods in Computational Biology.
MIT Press, 2004.
[12] T. Cover and J. Thomas. Elements of Information Theory. Wiley & Sons, New-York, 1991.
[13] N. Aronszajn. Theory of reproducing kernels. Transactions of the American Mathematical
Society, 68:337 ? 404, 1950.
| 2709 |@word determinant:1 kondor:1 briefly:1 norm:1 calculus:1 d2:1 grey:1 covariance:7 decomposition:1 homomorphism:1 initial:1 score:1 ecole:2 rkhs:3 tuned:1 existing:1 comparing:1 attracted:1 fn:2 numerical:1 shape:1 kyb:1 designed:1 discrimination:1 generative:1 selected:2 guess:1 merger:2 characterization:2 successive:1 simpler:1 mathematical:1 shtarkov:1 direct:1 prove:2 hermitian:1 expected:1 roughly:3 mpg:1 behavior:1 decreasing:1 equipped:1 considering:5 provided:1 bounded:5 didn:1 kind:1 string:2 finding:1 sensibility:1 every:2 ti:3 act:1 exactly:2 classifier:1 control:1 appear:1 positive:13 understood:1 local:1 consequence:1 bilinear:2 might:2 black:3 quantified:1 limited:2 range:4 averaged:1 practical:2 acknowledgment:1 testing:2 atomic:4 practice:2 recursive:1 definite:9 differs:1 orthonormalization:1 digit:6 ker:4 empirical:2 bell:1 thought:1 vert:5 word:1 numbered:2 protein:1 get:1 cannot:1 operator:6 context:2 writing:1 restriction:1 measurable:2 map:4 equivalent:3 www:1 fruitful:1 go:1 attention:1 graphically:1 independently:1 tjalkens:1 survey:1 splitting:1 toolbox1:1 estimator:1 spanned:1 regularize:1 orthonormal:3 his:1 laplace:3 qq:1 resp:2 suppose:3 exact:1 associate:2 element:9 recognition:1 database:3 hv:2 region:1 ensures:1 schoelkopf:1 decrease:2 highest:1 balanced:1 pd:2 cuturi:3 mine:2 trained:1 rewrite:1 basis:3 joint:1 represented:1 talk:1 train:1 describe:1 approached:1 choosing:1 exhaustive:1 jean:2 heuristic:1 whose:2 valued:7 solve:1 richer:1 amari:1 nagaoka:1 itself:1 final:1 associative:1 advantage:1 eigenvalue:5 sequence:2 propose:2 product:1 fr:2 relevant:1 kh:3 dirac:1 olkopf:1 qr:2 convergence:2 double:1 object:6 help:1 illustrate:2 depending:1 ij:1 quotient:1 predicted:1 kenji:1 direction:2 centered:2 require:1 suffices:1 f1:2 generalization:1 preliminary:1 proposition:5 summation:1 enumerated:2 marco:2 hold:1 around:1 considered:2 normal:1 exp:4 mapping:1 estimation:1 bag:2 weighted:2 fukumizu:1 mit:4 clearly:1 gaussian:8 always:3 aim:1 pn:5 avoid:1 ej:1 varying:1 casting:1 properly:1 likelihood:2 seeger:1 sense:3 am:1 dim:3 entire:1 pixel:8 issue:3 classification:5 among:4 dual:1 denoted:1 priori:1 ill:1 mutual:2 marginal:1 equal:1 once:1 having:2 extraction:1 sampling:1 biology:3 look:2 icml:1 representer:1 hint:1 randomly:3 composed:1 divergence:1 hvj:1 semigroups:3 argmax:3 geometry:1 jeffrey:1 recalling:1 attempt:1 alignment:1 yielding:2 admitting:1 tj:1 accurate:1 integral:4 respective:4 orthogonal:2 tree:1 euclidean:1 tsuda:1 renormalize:1 inconvenient:1 theoretical:1 instance:1 soft:1 cover:2 measuring:1 maximization:3 tg:1 deviation:1 neutral:1 subset:2 uniform:1 conducted:1 ensmp:2 answer:1 st:5 density:13 international:2 squared:1 aaai:1 american:1 supp:1 account:3 de:5 unordered:3 fontainebleau:2 gaussianity:2 coefficient:1 depends:2 vi:8 tion:1 view:1 doing:1 characterizes:1 reached:1 francis:1 shashua:1 contribution:1 variance:5 maximized:1 yield:4 correspond:1 generalize:1 weak:1 handwritten:3 bayesian:1 produced:1 rx:1 unaffected:1 definition:2 proof:3 associated:3 boil:2 couple:1 sampled:4 dataset:1 recall:3 knowledge:1 hilbert:2 formulation:1 though:2 furthermore:1 just:1 smola:1 correlation:1 sketch:1 hand:1 working:1 ei:5 aronszajn:1 brings:1 validity:1 concept:1 normalized:4 y2:4 hausdorff:2 xavier:1 hence:5 regularization:4 symmetric:3 semigroup:11 leibler:1 satisfactory:1 deal:1 width:4 uniquely:1 inferior:1 generalized:2 complete:3 fj:1 image:14 meaning:1 harmonic:1 recently:1 common:2 endowing:1 superior:1 exponentially:1 discussed:1 composition:1 cambridge:3 rd:1 tuning:1 trivially:1 consistency:1 pm:2 pq:1 dot:1 saigo:1 similarity:3 base:7 recent:1 perspective:1 systematical:1 certain:1 honor:2 binary:2 seen:8 captured:1 maximize:1 living:1 ing:1 characterized:1 calculation:1 cross:2 bach:2 regression:1 basic:1 kernel:70 represent:1 normalization:1 c1:1 addition:4 whereas:1 diagonalize:1 sch:1 specially:1 eigenfunctions:2 supposing:1 meaningfully:1 lafferty:1 jordan:1 call:2 noting:1 ideal:1 split:1 enough:2 xj:1 fit:2 identified:1 topology:2 restrict:1 idea:1 cn:1 tradeoff:1 expression:2 pca:1 york:1 useful:1 involve:1 eigenvectors:1 concentrated:1 svms:1 http:1 canonical:1 disjoint:1 vol:1 express:1 group:2 dominance:2 diffusion:2 v1:2 sum:2 semicharacters:2 angle:1 family:4 patch:1 scaling:2 radon:4 def:5 fold:5 encountered:1 nonnegative:1 yielded:1 vectorial:2 precisely:2 infinity:1 x2:5 argument:1 span:7 separable:1 kd:1 describes:1 son:1 character:1 making:2 b:1 christensen:1 multiplicity:1 taken:7 ln:7 equation:3 remains:2 previously:1 nonempty:1 letting:1 end:1 studying:2 endowed:6 apply:1 observe:2 alternative:1 schmidt:1 existence:1 original:3 thomas:1 remaining:1 ensure:2 saint:2 completed:1 establish:1 society:1 already:2 quantity:2 parametric:2 concentration:1 usual:2 diagonal:8 affinity:1 subspace:3 link:1 mapped:1 thank:1 concatenation:5 jakubowicz:1 ressel:1 tuebingen:1 reason:1 assuming:2 degenerated:2 pointwise:2 ellipsoid:1 providing:1 abelian:4 spider:3 negative:4 design:1 proper:1 perform:1 dispersion:2 willems:1 benchmark:1 finite:13 philippe:2 displayed:3 defining:2 extended:2 looking:1 y1:3 rn:4 reproducing:3 arbitrary:2 jebara:1 intensity:1 namely:8 paris:2 pair:1 toolbox:1 eigenfunction:2 beyond:1 below:2 pnto:1 emie:1 max:1 natural:2 warm:1 regularized:2 improve:1 carried:1 text:1 prior:4 geometric:1 review:1 law:4 expect:1 var:2 facing:1 versus:1 age:1 validation:2 editor:1 bypass:3 pi:1 repeat:1 allow:1 taking:2 curve:1 default:1 xn:3 valid:2 gram:4 preventing:1 forward:1 made:1 author:1 far:1 transaction:1 lebanon:1 kullback:1 dealing:1 ml:5 xi:8 spectrum:2 continuous:2 complex:3 rue:2 vj:10 main:1 dense:1 linearly:1 motivation:2 big:1 n2:2 allowed:1 x1:7 borel:2 definiteness:3 renormalization:2 wiley:1 fails:1 wish:1 explicit:1 exponential:4 lie:2 weighting:1 down:2 theorem:8 rk:2 bad:1 formula:3 er:1 list:16 pz:6 svm:4 bivariate:1 exists:1 mnist:3 adding:1 ci:8 margin:1 entropy:15 rayleigh:1 led:1 explore:1 likely:1 twentieth:1 akutsu:1 conveniently:1 expressed:1 contained:1 pinto:1 applies:1 springer:1 wolf:1 satisfies:1 extracted:1 ma:3 rbf:1 man:1 feasible:1 determined:2 infinite:1 typical:1 averaging:1 principal:2 called:2 total:2 experimental:1 berg:1 support:2 people:1 latter:7 constructive:1 tested:1 |
1,883 | 2,710 | Bayesian Regularization and Nonnegative
Deconvolution for Time Delay Estimation
Yuanqing Lin, Daniel D. Lee
GRASP Laboratory, Department of Electrical and System Engineering
University of Pennsylvania, Philadelphia, PA 19104
linyuanq, [email protected]
Abstract
Bayesian Regularization and Nonnegative Deconvolution (BRAND) is
proposed for estimating time delays of acoustic signals in reverberant
environments. Sparsity of the nonnegative filter coefficients is enforced
using an L1 -norm regularization. A probabilistic generative model is
used to simultaneously estimate the regularization parameters and filter
coefficients from the signal data. Iterative update rules are derived under
a Bayesian framework using the Expectation-Maximization procedure.
The resulting time delay estimation algorithm is demonstrated on noisy
acoustic data.
1 Introduction
Estimating the time difference of arrival is crucial for binaural acoustic sound source
localization[1]. A typical scenario is depicted in Fig. 1 where the azimuthal angle ? to
the sound source is determined by the difference in direct propagation times of the sound
to the two microphones. The standard signal processing algorithm for determining the
time delay between twoRsignals s(t) and x(t) relies upon computing the cross-correlation
function[2]: C(?t) = dt x(t)s(t ? ?t) and determining the time delay ?t that maximizes the cross-correlation. In the presence of uncorrelated white noise, this procedure is
equivalent to the optimal matched filter for detection of the time delayed signal.
However, a typical room environment is reverberant and the measured signal is contaminated with echoes from multiple paths as shown in Fig. 1. In this case, the cross-correlation
and related algorithms may not be optimal for estimating the time delays. An alternative
approach would be to estimate the multiple time delays as a linear deconvolution problem:
X
min kx(t) ?
?i s(t ? ?ti )k2
(1)
?
i
Unfortunately, this deconvolution can be ill-conditioned resulting in very noisy solutions
for the coefficients ?. Recently, we proposed incorporating nonnegativity constraints ? ?
0 in the deconvolution to overcome the ill-conditioned linear solutions [3]. The use of
these constraints is justified by acoustic models that describe the theoretical room impulse
response with nonnegative filter coeffients [4]. The resulting optimization problem can be
written as the nonnegative quadratic programming problem:
min kx ? S?k2
(2)
??0
s
?t?
?t1
?t2
?
x1
x2
d
Figure 1: The typical scenario of reverberant signal. x2 (t) comes from the direct path (?t2 ) and
echo paths(?tE ).
2.2
11
x 10
2
Phase Alignment Transform
Cross-correlation
800
a
600
1.8
400
1.6
200
1.4
0
1.2
-10
-5
0
5
time delay(T s )
10
Linear Deconvolution
200
-200
-10
1
c
100
0.8
0
0.6
-100
0.4
-200
0.2
-300
-10
-5
0
5
time delay(Ts )
10
b
-5
0
5
time delay(T s )
10
Nonnegative Deconvolution
d
0
-10
-5
0
5
time delay(Ts )
10
Figure 2: Time delay estimation of a speech signal with a) cross-correlation, b) phase alignment
transform, c) linear deconvolution, d) nonnegative deconvolution. The observed signal x(t) = s(t ?
Ts ) + 0.5s(t ? 8.75Ts ) contains an additional time-delayed echo. Ts is the sampling interval.
where x = {x(t1 ) x(t2 ) . . . x(tN )}T is a N ? 1 data vector, S = {s(t ? ?t1 ) s(t ?
?t2 ) . . . s(t ? ?tM )} is an N ? M matrix, and ? is a M ? 1 vector of nonnegative
coefficients.
Figure 2 compares the performance of cross-correlation, phase alignment transform(a generalized cross-correlation algorithm), linear deconvolution, and nonnegative deconvolution
for estimating the time delays in a clean speech signal containing an echo. From the structure of the estimated coefficients, it is clear that nonnegative deconvolution can successfully
discover the structure of the time delays present in the signal. However, in the presence of
large background noise, it may be necessary to regularize the nonnegative quadratic optimization to prevent overfitting. In this case, we propose using an L1 -norm regularization
to favor sparse solutions [5]:
X
?
min kx ? S?k2 + ?
?i
(3)
??0
i
? (?
? ? 0) describes the trade-off between fitting the observed
In this formula, the parameter ?
data and enforcing sparse solutions. The proper choice of this parameter may be crucial in
obtaining the optimal time delay estimates. In the rest of this manuscript, we introduce a
proper generative model for these regularization parameters and filter coefficients within a
probabilistic Bayesian framework. We show how these parameters can be efficiently determined using appropriate iterative estimates. We conclude by demonstrating and discussing
the performance of our algorithm on noisy acoustic signals in reverberant environments.
2 Bayesian regularization
Instead of arbitrarily setting values for the regularization parameters, we show how a
Bayesian framework can be used to automatically estimate the correct values from the
data. Bayesian regularization has previously been successfully applied to neural network
learning [6], model selection, and relevance vector machine (RVM) [7]. In these works, the
fitting coefficients are assumed to have Gaussian priors, which lead to an L2 -norm regularization. In our model, we use L1 -norm sparsity regularization, and Bayesian framework
will be used to optimally determine the appropriate regularization parameters.
Our probabilistic model assumes the observed data signal is generated by convolving the
source signal with a nonnegative filter describing the room impulse response. This signal
is then contaminated by additive Gaussian white noise with zero-mean and covariance ? 2 :
?
?
1
1
2
kx
?
S?k
.
(4)
P (x|S, ?, ? 2 ) =
exp
?
2? 2
(2?? 2 )N/2
To enforce sparseness in the filter coefficients ?, an exponential prior distribution is used.
This prior only has support in the nonnegative orthant and the sharpness of the distribution
is given by the regularization parameter ?:
P (?|?) = ?M exp{??
M
X
?i }, ? ? 0 .
(5)
i=1
In order to infer the optimal settings of the regularization parameters ? 2 and ?, Bayes rule
is used to maximize the posterior distribution:
P (?, ? 2 |x, S) =
P (x|?, ? 2 , S)P (?, ? 2 )
.
P (x|S)
(6)
Assuming that P (?, ? 2 ) is relatively flat [8], estimating ? 2 and ? is then equivalent to
maximizing the likelihood:
Z
?M
2
P (x|?, ? , S) =
d? exp[?F (?)]
(7)
(2?? 2 )N/2 ??0
where
1
(x ? S?)T (x ? S?) + ?eT ?
(8)
F (?) =
2? 2
T
and e = [1 1 . . . 1] .
Unfortunately, the integral in Eq. 7 cannot be directly maximized. Previous approaches
to Bayesian regularization have used iterative updates heuristically derived from selfconsistent fixed point equations. In our model, the following iterative update rules for
? and ? 2 can be derived using Expectation-Maximization:
Z
1
1
??
d? eT ?Q(?)
(9)
?
M ??0
Z
1
d? (x ? S?)T (x ? S?)Q(?)
(10)
? 2 ??
N ??0
where the expectations are taken over the distribution
Q(?) =
exp[?F (?)]
,
Z?
(11)
R
with normalization Z? = ??0 d? exp[?F (?)]. These updates have guaranteed convergence properties and can be intuitively understood as iteratively reestimating ? and ? 2
based upon appropriate expectations over the current estimate for Q(?).
2.1 Estimation of ?M L
The integrals in Eqs. 9?10 are dominated by ? ? ?M L where the most likely ?M L is
given by:
1
?M L = arg min 2 (x ? S?)T (x ? S?) + ?T ?.
(12)
??0 2?
This optimization is equivalent to the nonnegative quadratic programming problem in Eq. 3
? = ?? 2 . To efficiently compute ?M L , we have recently developed two distinct
with ?
methods for optimizing Eq. 12.
The first method is based upon a multiplicative update rule for nonnegative quadratic programming [9]. We first write the problem in the following form:
1
min ?T A? + bT ?,
??0 2
where A =
1 T
? 2 S S,
(13)
and b = ? ?12 ST x.
First, we decompose the matrix A = A+ ? A? into its positive and negative components
such that:
?
?
Aij if Aij > 0
0
if Aij ? 0
?
A+
A
(14)
=
=
ij
ij
0
if Aij ? 0
?Aij if Aij < 0
Then the following is an auxiliary function that upper bounds Eq. 13 [9]:
? = bT ? +
G(?, ?)
? i 2 1X ?
1 X (A+ ?)
?i ?j
?i ?
A ?
?i?
? j (1 + ln
).
2 i
?
?i
2 i,j ij
?
?i?
?j
(15)
Minimizing Eq. 15 yields the following iterative multiplicative rule with guaranteed convergence to ?M L :
p
?bi + b2i + 4(A+ ?)i (A? ?)i
.
(16)
?i ?? ?i
2(A+ ?)i
The iterative formula in Eq. 16 is used to efficiently compute a reasonable estimate for
?M L from an arbitrary initialization. However, its convergence is similar to other interior
point methods in that small components of ?M L will continually decrease but never equal
zero. In order to truly sparsify the solution, we employ an alternative method based upon
the simplex algorithm for linear programming.
Our other optimization method is based upon finding a solution ?M L that satistifies the
Karush-Kuhn-Tucker (KKT) conditions for Eq. 13:
A? + b = ?, ? ? 0, ? ? 0, ?i ?i = 0, i = 1, 2, . . . , M.
(17)
By introducing additional artificial
variables a, the KKT conditions can be transformed into
P
the linear optimization min i ai subject to the constraints:
a ? 0
(18)
? ? 0
(19)
? ? 0
(20)
A? ? ? + sign(?b)a = ?b
(21)
?i ?i = 0, i = 1, 2, . . . , M.
(22)
The only nonlinear constraint is the product ?i ?i = 0. However, this can be effectively
implemented in the simplex procedure by modifying the selection of the pivot element
to ensure that ?i and ?i are never both in the set of basic variables. With this simple
modification of the simplex algorithm, the optimal ?M L can be efficiently computed.
2.2 Approximation of Q(?)
Once the most likely ?M L has been determined, the simplest approach for estimating the
new ? and ? 2 in Eqs. 9?10 is to replace Q(?) ? ?(? ? ?M L ) in the integrals. Unfortunately, this simple approximation will cause ? and ? to diverge from bad initial estimates.
To overcome these difficulties, we use a slightly more sophisticated method of estimating
the expectations to properly consider variability in the distribution Q(?).
We first note that the solution ?M L of the nonnegative quadratic optimization in Eq. 12
naturally partitions the elements of the vector ? into two distinct subsets ?I and ?J , consisting of components i ? I such that (?M L )i = 0, and components j ? J such that
(?M L )j > 0, respectively. It will then be useful to approximate the distribution Q(?) as
the factored form:
Q(?) ? QI (?I )QJ (?J )
(23)
Consider the components ?J near the maximum likelihood solution ?M L . Among these
components, none of nonnegativity constraints are active, so it is reasonable to approximate
the distribution QJ (?J ) by the unconstrained Gaussian:
QJ (?J ) ? exp[?F (?J |?I = 0)]
(24)
ML
This Gaussian distribution has mean ?J and inverse covariance given by the submatrix
AJJ of A = ?12 ST S.
For the other components ?I , it is important to consider the nonnegativity constraints, since
L
?M
= 0 is on the boundary of the distribution. We can represent QI (?I ) with the first
I
two order Tyler expansion:
?F
1
)|?M L ]TI ?I ? ?TI AII ?I )},
QI (?I ) ? exp{?[(
??
2
1
ML
T
? exp[?(A?
+ b)I ?I ? ?TI AII ?I ]
2
?I ? 0.
(25)
? I (?I ) so that the
QI (?I ) is then approximated with factorial exponential distribution Q
integrals in Eqs. 9?10 can be easily evaluated.
Y 1
? I (?I ) =
e??i /?i , ?I ? 0
(26)
Q
?i
i?I
which has support only for nonnegative ?I ? 0. The mean-field parameters ? are optimally obtained by minimizing the KL-divergence:
Z
?
? I (?I ) ln QI (?I ) .
(27)
min
d?I Q
??0 ?I ?0
QI (?I )
This integral can easily be computed in terms of the parameters ? and yields the minimization:
X
? T ? + 1 ?T A?,
?
(28)
min ?
ln ?i + b
I
??0
2
i?I
? I = (A?M L + b)I , A
? = AII + diag(AII ). To solve this minimization problem,
where b
we use an auxiliary function for Eq. 28 similar to the auxiliary function for nonnegative
quadratic programming:
? =?
G(?, ?)
X
i?I
?T ? + 1
ln ?i + b
I
2
X (A
? + ?)
? i
i?I
?
?i
?2i ?
1 X ??
?i ?j
Aij ?
?i ?
?j (1 + ln
), (29)
2
?
?i ?
?j
i,j?I
? =A
? + ?A
? ? is the decomposition of A
? into its positive and negative components.
where A
Minimization of this auxiliary function yields the following multiplicative update rules for
?i :
q
? + ?)i [(A
? ? ?)i + 1 ]
??bi + ?b2i + 4(A
?i
.
(30)
?i ?? ?i
+
?
2(A ?)i
These iterations are then guaranteed to converge to the optimal mean-field parameters for
the distribution QI (?I ).
? I (?I )QJ (?J ), the expectations in Eqs. 9?10 can be
Given the factorized approximation Q
analytically calculated. The mean value of ? under this distribution is given by:
?
?iM L if i ? J
(31)
?
?i =
?i
if i ? I
and its covariance C is:
?
Cij =
(AJJ ?1 )ij
?2i ?ij
if i, j ? J
otherwise
(32)
The update rules for ? and ? 2 are then given by:
?
??
?2
??
M
P
?i
i?
1
? T (x ? S?)
? + Tr(ST SC)]
[(x ? S?)
N
(33)
(34)
To summarize, the complete algorithm consists of the following steps:
1. Initialize ? and ? 2 .
2. Determine ?M L by solving the nonnegative quadratic programming in Eq. 12.
? I (?I )QJ (?J ) by solving the mean field
3. Approximate the distribution Q(?) ? Q
?I .
equations for ? in Q
? and covariance C for this distribution.
4. Calculate the mean ?
5. Reestimate regularization parameters ? and ? 2 using Eqs. 33?34.
6. Go back to Step 2 until convergence.
3
1.8
10
1.6
-40dB
40dB
-20dB
20dB
- 5dB
5dB
2
10
1.4
0
10
1
22
??
?/?
?/?
1
10
-40dB
40dB
-20dB
20dB
- 5dB
5dB
1.2
0.8
1
-1
10
0.6
2
-2
10
0.4
3
-3
10
0.2
0
4
-4
0
5
10
Iteration Number
15
20
10
0
5
10
Iteration Number
15
20
Figure 3: Iterative estimation of ? (M/? in the figure, indicating the reverberation level) and ? 2
when x(t) is contaminated by background white noise at -5 dB, -20 dB, and -40 dB levels. The
horizontal dotted lines indicate the true levels.
3 Results
We illustrate the performance of our algorithm in estimating the regularization parameters
as well as the nonnegative filter coefficients of a speech source signal s(t). The observed
signal x(t) is simulated by a time-delayed version of the source signal mixed with an echo
along with additive Gaussian white noise ?(t):
x(t) = s(t ? Ts ) + 0.5s(t ? 16.5Ts ) + ?(t).
(35)
We compare the results of the algorithm as the noise level is changed. Fig. 3 shows the
convergence of the estimates for ? and ? 2 as the noise level is varied between -5 dB and
-40 dB. There is rapid convergence of both parameters even with bad initial estimates. The
resulting value of the ? 2 parameter is very close to the true noise level. Additionally, the
estimated ? parameter is inversely related to the reverberation level of the environment,
given by the sum of the true filter coefficients.
Fig. 4 demonstrates the importance of correctly determining the regularization parameters
in estimating the time delay structure in the presence of noise. Using the Bayesian regularization procedure, the resulting estimate for ?M L correctly models the direct path time
delay as well as the secondary echo. However, if the regularization parameters are manually set incorrectly to over-sparsify the solution, the resulting estimates for the time delays
may be quite inaccurate.
4 Discussion
In summary, we propose using a Bayesian framework to automatically regularize nonnegative deconvolutions for estimating time delays in acoustic signals. We present two methods
for efficiently solving the resulting nonnegative quadratic programming problem. We also
derive an iterative algorithm from Expectation-Maximization to estimate the regularization
parameters. We show how these iterative updates can simulataneously estimate the timedelay structure in the signal, as well as the background noise level and reverberation level
of the room. Our results indicate that the algorithm is able to quickly converge to an optimal
solution, even with bad initial estimates. Preliminary tests with an acoustic robotic platform
indicate that these algorithms can successfully be implemented on a real-time system.
?=50, ?2 = 0.12
1
a
0.8
?
0.6
0.4
0.2
0
-20
-15
-10
-5
0
5
10
15
20
5
10
15
20
Time delay (Ts )
2
?? =200
1
b
0.8
?
0.6
0.4
0.2
0
-20
-15
-10
-5
0
Time delay (Ts )
Figure 4: Estimated time delay structure from ?M L with different regularizations: a) Bayesian
regularization, b) manually set regularization. Dotted lines indicate the true positions of the time
delays.
We are currently working to extend the algorithm to the situation where the source signal
needs to also be estimated. In this case, priors for the source signal are used to regularize
the source estimates. These priors are similar to those used for blind source separation. We
are investigating algorithms that can simultaneously estimate the hyperparameters for these
priors in addition to the other parameters within a consistent Bayesian framework.
References
[1] E. Ben-Reuven and Y. Singer, ?Discriminative binaural sound localization,? in Advances in Neural Information Processing Systems, S. T. Suzanna Becker and K. Obermayer, Eds., vol. 15. The
MIT Press, 2002.
[2] C. H. Knapp and G. C. Carter., ?The generalized correlation method for estimation of time delay,?
IEEE Transactions on ASSP, vol. 24, no. 4, pp. 320?327, 1976.
[3] Y. Lin, D. D. Lee, and L. K. Saul, ?Nonnegative deconvolution for time of arrival estimation,? in
ICASSP, 2004.
[4] J. B. Allen and D. A. Berkley, ?Image method for efficient simulating small-room acoustics,? J.
Acoust. Soc. Am., vol. 65, pp. 943?950, 1979.
[5] B. Olshausen and D. Field, ?Emergence of simple-cell receptive field properties by learning a
sparse code for nature images,? Nature, vol. 381, pp. 607?609, 1996.
[6] D. Foresee and M. Hagan, ?Gauss- Newton approximation to Bayesian regularization,? in Proceedings of the 1997 International Joint Conference on Neural Networks, 1997, pp. 1930?1935.
[7] M. E. Tipping, ?Sparse Bayesian learning and the relevance vector machine,? Journal of Machine
Learning Research, vol. 1, pp. 211?244, 2001.
[8] D. MacKay, ?Bayesian interpolation,? Neural Computation, vol. 4, pp. 415?447, 1992.
[9] F. Sha, L. K. Saul, and D. Lee, ?Multiplicative updates for nonnegative quadratic programming in support vector machines,? in Advances in Neural Information Processing Systems, S. T.
Suzanna Becker and K. Obermayer, Eds., vol. 15. The MIT Press, 2002.
| 2710 |@word version:1 norm:4 deconvolutions:1 heuristically:1 azimuthal:1 ajj:2 covariance:4 decomposition:1 tr:1 initial:3 contains:1 daniel:1 current:1 written:1 additive:2 partition:1 update:9 generative:2 along:1 direct:3 consists:1 fitting:2 introduce:1 upenn:1 rapid:1 automatically:2 estimating:10 matched:1 discover:1 maximizes:1 factorized:1 developed:1 acoust:1 finding:1 ti:4 k2:3 demonstrates:1 continually:1 t1:3 positive:2 engineering:1 understood:1 path:4 interpolation:1 initialization:1 bi:2 procedure:4 cannot:1 interior:1 selection:2 close:1 equivalent:3 demonstrated:1 maximizing:1 go:1 sharpness:1 suzanna:2 factored:1 rule:7 regularize:3 programming:8 pa:1 element:2 approximated:1 hagan:1 observed:4 electrical:1 calculate:1 trade:1 decrease:1 environment:4 solving:3 localization:2 upon:5 binaural:2 easily:2 joint:1 icassp:1 aii:4 distinct:2 describe:1 artificial:1 sc:1 quite:1 solve:1 otherwise:1 favor:1 transform:3 noisy:3 echo:6 emergence:1 propose:2 product:1 convergence:6 sea:1 ben:1 illustrate:1 derive:1 measured:1 ij:5 eq:15 soc:1 auxiliary:4 implemented:2 come:1 indicate:4 kuhn:1 correct:1 filter:9 modifying:1 karush:1 decompose:1 preliminary:1 im:1 exp:8 tyler:1 estimation:7 currently:1 rvm:1 successfully:3 minimization:3 mit:2 gaussian:5 sparsify:2 derived:3 properly:1 likelihood:2 am:1 inaccurate:1 bt:2 transformed:1 arg:1 among:1 ill:2 platform:1 initialize:1 mackay:1 equal:1 once:1 never:2 field:5 sampling:1 manually:2 simplex:3 contaminated:3 t2:4 employ:1 simultaneously:2 divergence:1 delayed:3 phase:3 consisting:1 detection:1 grasp:1 alignment:3 truly:1 integral:5 necessary:1 theoretical:1 maximization:3 introducing:1 subset:1 delay:24 optimally:2 reuven:1 st:3 international:1 lee:3 probabilistic:3 off:1 diverge:1 quickly:1 containing:1 convolving:1 coefficient:10 blind:1 b2i:2 multiplicative:4 bayes:1 efficiently:5 maximized:1 yield:3 bayesian:16 none:1 reestimate:1 ed:2 pp:6 tucker:1 naturally:1 sophisticated:1 back:1 manuscript:1 dt:1 tipping:1 response:2 evaluated:1 foresee:1 correlation:8 until:1 working:1 horizontal:1 nonlinear:1 propagation:1 impulse:2 olshausen:1 true:4 regularization:25 analytically:1 laboratory:1 iteratively:1 white:4 timedelay:1 generalized:2 complete:1 tn:1 l1:3 allen:1 image:2 recently:2 extend:1 ai:1 unconstrained:1 selfconsistent:1 berkley:1 posterior:1 optimizing:1 scenario:2 arbitrarily:1 discussing:1 additional:2 determine:2 maximize:1 converge:2 signal:21 multiple:2 sound:4 infer:1 cross:7 lin:2 qi:7 basic:1 expectation:7 iteration:3 normalization:1 represent:1 cell:1 justified:1 background:3 addition:1 interval:1 source:9 crucial:2 rest:1 subject:1 db:17 near:1 presence:3 reverberant:4 pennsylvania:1 tm:1 pivot:1 qj:5 becker:2 speech:3 cause:1 useful:1 clear:1 factorial:1 carter:1 simplest:1 dotted:2 sign:1 estimated:4 correctly:2 write:1 vol:7 demonstrating:1 prevent:1 clean:1 sum:1 enforced:1 angle:1 inverse:1 reasonable:2 separation:1 submatrix:1 bound:1 guaranteed:3 quadratic:9 nonnegative:24 constraint:6 x2:2 flat:1 dominated:1 min:8 relatively:1 department:1 describes:1 slightly:1 modification:1 intuitively:1 taken:1 ln:5 equation:2 previously:1 describing:1 singer:1 appropriate:3 enforce:1 simulating:1 alternative:2 assumes:1 ensure:1 newton:1 receptive:1 sha:1 obermayer:2 simulated:1 enforcing:1 yuanqing:1 assuming:1 code:1 minimizing:2 unfortunately:3 cij:1 negative:2 reverberation:3 proper:2 upper:1 t:9 orthant:1 incorrectly:1 situation:1 variability:1 assp:1 varied:1 arbitrary:1 kl:1 acoustic:8 able:1 sparsity:2 summarize:1 difficulty:1 inversely:1 philadelphia:1 prior:6 l2:1 knapp:1 determining:3 mixed:1 reestimating:1 consistent:1 uncorrelated:1 changed:1 summary:1 aij:7 saul:2 sparse:4 overcome:2 boundary:1 calculated:1 transaction:1 approximate:3 ml:2 overfitting:1 kkt:2 active:1 robotic:1 investigating:1 conclude:1 assumed:1 discriminative:1 iterative:9 additionally:1 nature:2 obtaining:1 expansion:1 diag:1 noise:10 hyperparameters:1 arrival:2 x1:1 fig:4 ddlee:1 position:1 nonnegativity:3 exponential:2 formula:2 bad:3 deconvolution:13 incorporating:1 effectively:1 importance:1 te:1 conditioned:2 sparseness:1 kx:4 depicted:1 likely:2 relies:1 room:5 replace:1 typical:3 determined:3 microphone:1 secondary:1 gauss:1 brand:1 indicating:1 support:3 relevance:2 |
1,884 | 2,711 | Similarity and discrimination in classical
conditioning: A latent variable account
Aaron C. Courville*1,3 , Nathaniel D. Daw4 and David S. Touretzky2,3
1
Robotics Institute, 2 Computer Science Department,
3
Center for the Neural Basis of Cognition,
Carnegie Mellon University, Pittsburgh, PA 15213
4
Gatsby Computational Neuroscience Unit, University College London
{aaronc,dst}@cs.cmu.edu; [email protected]
Abstract
We propose a probabilistic, generative account of configural learning
phenomena in classical conditioning. Configural learning experiments
probe how animals discriminate and generalize between patterns of simultaneously presented stimuli (such as tones and lights) that are differentially predictive of reinforcement. Previous models of these issues
have been successful more on a phenomenological than an explanatory
level: they reproduce experimental findings but, lacking formal foundations, provide scant basis for understanding why animals behave as they
do. We present a theory that clarifies seemingly arbitrary aspects of previous models while also capturing a broader set of data. Key patterns
of data, e.g. concerning animals? readiness to distinguish patterns with
varying degrees of overlap, are shown to follow from statistical inference.
1 Introduction
Classical conditioning experiments probe how organisms learn to predict significant events
such as the receipt of food or shock. While there is a history of detailed quantitative theories about these experiments, only recently has there been a sustained attempt to understand
them in terms of sound statistical prediction [1]. A statistical foundation helps to identify
key theoretical issues (such as uncertainty) underlying these experiments, to explain otherwise puzzling results, and to connect these behavioral theories with theories of neural
computation, which are also increasingly framed in statistical terms.
A cluster of issues that has received great experimental and theoretical attention in conditioning ? but not yet from a statistically grounded perspective ? concerns discrimination
and generalization between patterns of sensory input. Historically, these issues arose in the
context of nonlinear discriminations, such as the XOR problem (in which, e.g., a light and
a tone each predict shock when presented alone, but not together). While animals can learn
such a discrimination, the seminal model of Rescorla and Wagner [2] cannot, since it assumes that the prediction is linear in the stimuli. Traditionally, this problem was solved by
introducing extra discriminative features to the model?s input (known as ?configural units,?
since they detect conjunctions of stimuli such as tone plus light), rendering the augmented
problem linearly solvable [3]. On this foundation rests a wealth of work probing how ani-
mals learn and predict given compounds of stimuli. Here, we reinterpret these issues from
a Bayesian perspective.
Previous work posits an informal division (or perhaps a spectrum) between ?elemental?
and ?configural? approaches to stimulus patterns, distinguished by whether a compound?s
association with reinforcement is derived from its individual stimuli (lights, tones), or rests
collectively in the full compound (light and tone together). The prototypical elemental
model is the original Rescorla?Wagner model, without configural units, in which the aggregate prediction is linear in the elements. The standard configural model is that of Pearce
[4], in which responding to a compound is determined by previous experience with that
and other similar compounds, through a process of generalization and weighted averaging.
Both theories match an impressive range of experimental data, but each is refuted by some
experiments that the other captures. It is not clear how to move beyond this stalemate.
Because the theories lack formal foundations, their details ? particularly those on which
they differ ? are ad-hoc and poorly understood. For instance, what circumstances justify
the introduction of a new configural unit, and what should be the form of generalization
between compounds?
Here we leverage our Bayesian theory of conditioning [5] to shed new light on these issues.
Our model differs from traditional ones in a number of ways. Notably, analogizing conditioning to classification, we take a generative rather than a discriminative approach. That
is, we assume animals are modeling their complete sensory experience (lights, tones, and
shocks) rather than only the chance of shock conditioned on lights and tones. We assume
that stimuli are correlated with each other, and with reinforcement, through shared latent
variables. Because a latent variable can trigger multiple events, these causes play a role
akin to configural units in previous theories, but offer stronger normative guidance. Questions about generalization (what is the probability that a latent variable is active given a
particular constellation of inputs) are seen as standard statistical inference; questions about
model structure (how many ?configural units? should there be and with what constellations
of stimuli are they associated) are answerable using Bayesian model averaging, which we
have suggested animals can approximate [5]. Such inferences also determine whether an
animal?s experience on a trial is best explained by multiple causes interacting additively, in
the style of Rescorla?Wagner, or by a single cause triggering multiple events like one of
Pearce?s configural units. This allows our theory to capture patterns of data that seem to
favor each of its predecessors.
Our theory is meant to shed light on the normative reasons why animals behave as they do,
rather than on how they might carry out computations like those we describe. In practice,
the inferences we discuss can be computed only approximately, and we intend no claim that
animals are using the same approximations to them as we are. More mechanistic models,
such as Pearce?s, can broadly be viewed as plausible implementations for approximating
some aspects of our more general framework.
2 Theories of Learning with Compound Stimuli
Classical conditioning experiments probe animals? anticipation of a reinforcer R such as
food or footshock, given the presentation of initially neutral stimuli such as lights and tones.
Expectation is assessed via reflexive conditioned responses such as salivation or freezing,
which are thought to reveal animals? predictions of reinforcement. By studying responding
as a function of the pattern of previous reinforcer / stimulus pairings, the experiments assess
learning. To describe a conditioning task abstractly, we use capital letters for the stimuli
and + and ? to indicate whether they are reinforced. For instance, the XOR task can be
written as A+, B+, AB?, where AB? denotes simultaneous presentation of both stimuli
unreinforced. Typically, each type of trial is delivered repeatedly, and the development of
responding is assessed.
We now describe the treatment of compound stimuli in the models of Rescorla and Wagner
[2] and Pearce [4]. In both models, the set of stimuli present on a trial is converted into
an input vector x. The strength of the conditioned response is modeled as proportional to
a prediction of reinforcement v = x ? w, the dot product between the input and a weight
vector. Finally, one or more weights are updated proportionally to the mismatch r ? v
between observed and predicted reinforcement.
For both theories, x includes an element (or ?unit?) corresponding to each individual stimulus.1 In Pearce?s model, and in augmented ?added elements? versions of the Rescorla?
Wagner model [3], additional ?configural? units are also included, corresponding to conjunctions of stimuli. In particular, it is assumed that a unique configural unit is added for
each stimulus compound observed, such as ABC. Note that this assumption is both arbitrary (e.g. we might very well include elements for subcompounds such as AB) and
unrealistic (given the profusion of uncontrolled stimuli simultaneously present in a real
experiment).
The theories differ as to how they apportion activation over x and learning over w. In the
Rescorla?Wagner model, the input vector is binary: xi = 1 if the ith stimulus (or an exactly
matching compound) is present, 0 otherwise. For learning, the weight corresponding to
each active input is updated. The Pearce model instead spreads graded activation over x,
based on a measure of similarity between the observed stimulus compound (or element)
and the compounds represented by the model?s configural units. In particular, if we denote
the number of stimulus elements present in an observed stimulus pattern a as size(a), and
in the pattern represented by the ith configural unit as size(i), then the activation of unit
i by pattern a is given by xi = size(overlap(a, i))2 /(size(a) ? size(i)). The learning
phase updates only the weight corresponding to the configural unit that exactly matches
the observed stimulus configuration.
As neither scheme has much formal basis, there seems to be no theoretical reason to prefer
one over the other, nor over any other ad-hoc recipe for apportioning representation and
learning. Empirical considerations also provide ambivalent guidance, as we discuss next.
3 Data on Learning with Compound Stimuli
Both the elemental and configural models reproduce a number of well known experimental
phenomena. Here we review several basic patterns of results. Notably, each theory has
a set of experiments that seems to support it over the other. Later, we will show that our
normative theory accounts for all of these results.
Overshadowing When a pair of stimuli AB+ is reinforced together, then tested separately, responding to either individual stimulus is often attenuated compared to a control in
which the stimulus is trained alone (A+). Both models reproduce overshadowing, though
Rescorla?Wagner incorrectly predicts that it takes at least two AB+ pairings to materialize.
Summation The converse of overshadowing is summation: when two stimuli are individually reinforced, then tested together, there is often a greater response to the pair than to
either element alone. In a recent variation by Rescorla [6], animals were trained on a pair
of compounds AB+ and CD+, then responses were measured to the trained compounds,
the individual elements A, B, etc., and the novel transfer compounds AD and BC. The
strongest response was elicited by the trained compounds. The transfer compounds elicited
a moderate response, and the individual stimuli produced the weakest responding.
1
In Pearce?s presentation of his model, these units are added only after elements are observed
alone. We include them initially, which does not affect the model?s behavior, to stress similarity with
the Rescorla-Wagner model.
The added elements Rescorla?Wagner model predicts this result due to the linear summation of the influences of all the units (A through D, AB, and CD ? note that the added
configural units are crucial). However, because of the normalization term in the generalization rule, Pearce?s model often predicts no summation. Here it predicts equal responding
to the individual stimuli and to the transfer compounds. There is controversy as to whether
the model can realistically be reconciled with summation effects [4, 7], but on the whole,
these phenomena seem more parsimoniously explained with an elemental account.
Overlap A large number of experiments (see [4] for a review) demonstrate that the more
elements shared by two compounds, the longer it takes animals to learn to discriminate
between them. Though this may seem intuitive, elemental theories predict the opposite.
In one example, Redhead and Pearce [8] presented subjects with the patterns A+, BC+
reinforced and ABC? unreinforced. Differential responding between A and ABC was
achieved in fewer trials than that between BC and ABC.
Pearce?s configural theory predicts this result because the extra overlap between BC and
ABC (compared to A vs. ABC) causes each compound to activate the other?s configural
unit more strongly. Thus, larger weights are required to produce a differentiated prediction.
Rescorla?Wagner predicts the opposite result, because compounds with more elements, e.g.
BC, accumulate more learning on each trial.
4 A latent variable model of stimulus generalization
In this section we present a generative model of how stimuli and reinforcers are jointly
delivered. We will show how the model may be used to estimate the conditional probability of reinforcement (the quantity we assume drives animals? responding) given some
pattern of observed stimuli. The theory is based on the one we presented in [5], and casts
conditioning as inference over a set of sigmoid belief networks. Our goal here is to use this
formalism to explain configural learning phenomena.
4.1 A Sigmoid Belief Network Model of Conditioning
Consider a vector of random variables S representing stimuli on a trial, with the jth stimulus present when Sj = 1 and absent when Sj = 0. One element of S is distinguished as the
reinforcer R; the remainder (lights and tones) is denoted as Stim. We encode the correlations between all stimuli (including the reinforcer) through common connections to a vector
of latent variables, or causes, x where xi ? {0, 1}. According to the generative process,
on each trial the state of the latent variables is determined by independent Bernoulli draws
(each latent variable has a weight determining its chance of activation [5]). The probability
of stimulus j being present is then determined by its relationship to the latent variables:
(j) T
P (Sj | m, wm , x) = (1 + exp(?(wm
) x ? wbias ))?1 ,
(1)
(j)
where the weight vector wm encodes the connection strengths between x and Sj for the
model structure m. The bias weight wbias is fixed at ?6, ensuring that spontaneous events
are rare. Some examples of the type of network structure under consideration are shown as
graphical models in Figure 1(c)?(d) and Figure 2(c)?(e).
We assume animals learn about the model structure itself, analogous to the experiencedependent introduction of configural units in previous theories. In our theory, animals
use experience to infer which network structures (from a set of candidates) and weights
likely produced the observed stimuli and reinforcers. These in turn determine predictions
of future reinforcement. Details of this inference are laid out below.
4.2 Generalization: inference over latent variables
Generalization between observed stimulus patterns is a key aspect of previous models. We
now describe how generalization arises in our theory.
Given a particular belief net structure m, weights wm , and previous conditioning experience D, the probability of reinforcement R given observed stimuli Stim can be computed
by integrating over the possible settings x of the latent variables:
X
P (R | Stim, m, wm , D) =
P (R | m, wm , x)P (x | Stim, m, wm , D)
(2)
x
The first term is given by Equation 1. By Bayes? rule, the second term weighs particular
settings of the hidden causes proportionally to the likelihood that they would give rise to the
observed stimuli. This process is a counterpart to Pearce?s generalization rule for configural
units. Unlike Pearce?s rule, inference over x considers settings of the individual causes xi
jointly (allowing for explaining away effects) and incorporates prior probabilities over each
cause?s activation. Nevertheless, the new rule broadly resembles its predecessor in that a
cause is judged likely to be active (and contributes to predicting R) if the constellation of
stimuli it predicts is similar to what is observed.
4.3 Learning to discriminate: inference over models
We treat the model weights wm and the model structure m as uncertain quantities subject
to standard Bayesian inference. We assume that, given a model structure, the weights are
mutually independent a priori and each distributed according to a Laplace distribution.2
Conditioning on the data D produces a posterior distribution over the weights, over which
we integrate to predict R:
Z
P (R | Stim, m, D) = P (R | Stim, m, wm , D)P (wm | m, D)dwm
(3)
Uncertainty over model structure is handled analogously. Integrating over posterior model
uncertainty we arrive at the prediction of reinforcement:
X
P (R | Stim, D) =
P (R | Stim, m, D)P (m | D),
(4)
m
where P (m | D) ? P (D | m)P (m) and the marginal likelihood P (D | m) is computed
similarly to equation 3, by integration over the weights. The prior over models, P (m) is
expressed as a distribution over nx , the number of latent variables, and over
Qnlxi , the number
P (li ). We
of links between the stimuli and each latent variable: P (m) = P (nx ) i=1
assume that P (nx ) and each P (li ) are given by geometric distributions (param. = 0.1),
renormalized to sum to unity over the maximum of 5 latents and 5 stimuli. This prior
reflects a bias against complex models. The marginal likelihood term also favors simplicity,
due to the automatic Occam?s razor (see [5]). For our simulations, we approximately
evaluated Equation 4 using reversible-jump Markov Chain Monte Carlo (see [5] for details).
Progressively conditioning on experience to resolve prior uncertainty in the weights and
model structure produces a gradual change in predictions akin to the incremental learning
rules of previous models. The extent to which a particular model structure m participates
in predicting R in Equation 4 is, by Bayes? rule, proportional to its prior probability, P (m),
and to the extent that it explains the data, P (D | m). Thus a prior preference for simpler
models competes against better data fidelity for more complex models. As data accumulate,
1 ?|y??|/b
The Laplace distribution is given by f (y) = 2b
e
. In our simulations ? = 0 and b = 2.
As a prior, it encodes a bias for sparsity consistent with a preference for simpler model structures.
2
1
1
0.8
0.8
P(R | Probe,D )
P(R | Probe,D )
x1
0.6
0.4
0.2
0
0.6
A
0.4
AB
Control
Probe Stimulus
(a) Overshadowing
0
R
(c) Overshadowing
x
0.2
A
B
x
1
Trained
Transfer
Probe Stimulus
(b) Summation
2
Element
A
B
C
D
R
(d) Summation
Figure 1: Results of MCMC simulation. (a) Overshadowing (AB+): the predicted probability of
reinforcement in response to presentations of the element A, the compound AB, and an individually
trained control element (A+). (b) Summation experiment (AB+, CD+): the predicted probability
of reinforcement in response to separate presentations of the trained compounds (AB, CD), the
transfer compounds (AD, BC) and the elements (A, B, etc.). (c) Depiction of the MAP model
structure after overshadowing training. (d) The MAP model structure after AB+ CD+ training.
the balance shifts toward the latter, and predictions become more accurate. Analogously,
weights are small a priori but can grow with experience.
Together with the generalization effects discussed above, these inference effects explain
why animals can learn more readily to discriminate stimulus compounds that have less
overlap. Key to the discrimination is inferring that different compounds are produced by
separate latent variables; the more the compounds overlap, the more accurately will the
data be approximated by a model with a single latent variable (preferred a priori), which
biases the complexity-fidelity tradeoff toward simplicity and retards acquisition.
5 Results
Overshadowing Overshadowing exemplifies our account of between-compound generalization; our model?s performance is illustrated in Figure 1(a). After 5 AB+ pairings,
the network with highest posterior probability, depicted in (c), contains one latent variable correlated with both stimuli and the reinforcer. Consistent with experimental results,
testing on A produces attenuated responding. This is because predicting whether A is reinforced requires balancing the relative probabilities of two unlikely events: that the stimulus
occurred spontaneously (with x1 inactive), versus that it was caused by x1 being active,
but that B uncharacteristically failed to occur (this probability measures generalization between the patterns A and AB). Overall, this tradeoff decreases the chance that x1 is active,
suppressing the prediction of reinforcement relative to the control treatment, where A is
reinforced in isolation (A+). Unlike the Rescorla?Wagner model, ours correctly predicts
that overshadowing can occur after even a single AB+ presentation.
Summation Figure 1(b) shows our model?s performance on Rescorla?s AB+ CD+ summation and transfer experiment [6], which is one of several summation experiments our
model explains. Compounds were reinforced 10 times. Consistent with experimental findings, the model predicts greatest responding to the trained compounds (AB, CD), moderate responding to transfer compounds (AD, BC), and least responding to the elements (A,
B, etc.). The maximum a posteriori (MAP) model structure (Figure 1(d)) mimics the training compounds, with one latent variable connected to A, B, and R and another connected
to C, D, and R. The results follow from a combination of generalization and additivity.
The training compounds activate one latent variable strongly; the transfer compounds acti-
3.5
Average number of latent variables
1
0.8
0.6
P(R | A, D )
P(R | B,C, D )
P(R | A,B,C, D )
0.4
0.2
0
0
5
10
15
20
Trial Blocks
25
1
A
B
R
(c) 4 Trial Blocks
A
1
0.5
5
10
15
20
Trial Blocks
25
30
(b) Model growth with training
x
x
B
C
1
C
2
1.5
0
0
30
(a) Overlap: learning curves
x
3
2.5
x
2
x
1
R
(d) 10 Trial Blocks
A
x
2
B
3
C
R
(e) 20 Trial Blocks
Figure 2: Summary of MCMC simulation results on the A+, BC+, ABC? experiment. The
estimated error due to MCMC sampling is small and not shown. (a) Learning curves showing the
predicted probability of reinforcement in response to separate presentations of A, BC, and ABC
as a function of number of trial blocks. (b) The average number of latent variables over the 10000
MCMC sample models. (c) - (e) Representations of MAP model structures after training with 4, 10,
and 20 trial blocks (edge widths represent mean weight strength).
vate both latents weakly (together additively influencing the probability of reinforcement);
the elements weakly activate only a single latent variable.
Overlap Figure 2(a) shows the model?s learning curves from the overlapping compound
experiment, A+, BC+, ABC?. Each trial block contains one trial of each type. The
model correctly predicts faster discrimination between A and ABC than between BC and
ABC. This pattern results from progressive increase in the number of inferred latent variables (b). Early in training, probability density concentrates on small models with a single
latent variable correlating all stimuli and the reinforcer (c). After more trials, models with
two latent variables become more probable, one correlating A and R and the other correlating B and C with both A and R, attempting to capture both BC+ and ABC? trial
types. (d). With further training, the most likely models are those with three latents, each
encoding one trial type (e). Our theory captures many similar experiments demonstrating
the difficulty of discriminating overlapping compounds.
6 Discussion
The configural unit is an ad-hoc device that nonetheless plays a key role in previous experimental and theoretical work in conditioning. Its inclusion in models like that of Rescorla?
Wagner invites a number of questions. Which configurations should be represented? How
should activation and learning be apportioned between them? These issues are contentious,
admitting no clear answer, precisely because of the arbitrary nature of the device. We
have shown how a latent variable correlated with a constellation of stimuli provides a well
founded counterpart to the configural unit, and how a range of experimental phenomena
concerning similarity and discrimination can be accounted for with the assumption that
animals are carrying out inference about these variables. While data exist that tend to fa-
vor each of the two major previous models of configural learning over the other, the new
model accounts for the full pattern, balancing the strengths of both theories. Our theory
also improves on its predecessors in other ways; for instance, because it includes learning
about stimulus interrelationships it can explain second-order conditioning [5], which is not
addressed by either the Pearce or the Rescorla?Wagner accounts.
Of course, many issues remain. A full account of summation phenomena, in particular, is
beyond the scope of the present model. We treat reinforcer delivery as binary and model
a limited, saturating, summation in probabilities. However, realistic summation almost
certainly concerns reinforcement magnitudes as well (see, for example, [9]), and our model
would need to be augmented to address them. Because we have assumed that trials are IID,
the model cannot yet account for effects of trial ordering (e.g. the difference between partial
reinforcement and extinction). These could be addressed by incorporating dynamics into
the generative model, so that inference requires tracking the changing model parameters.
Also for future work is exploring how different priors might give rise to different behavior.
An advantage of Bayesian modeling is that because the free parameters are formulated as
priors, they represent concrete assertions about the world (e.g. how often particular kinds
of events occur), and can thus be constrained and even experimentally manipulated.
We have focused only on two previous models and only on animal behavioral experiments.
Issues of similarity and discrimination are also studied in the rather different setting of human category judgments, where Bayesian generative approaches have also proved useful
[10]. There is also a tradition of more neurophysiological models of the hippocampal substrates of configural learning [11, 12]. Given the large body of theory and experiment on
these issues, this seems a promising direction for future work connecting our behavioral
theory with neurophysiological ones. In one of the hippocampal theories, Gluck and Myers [12] augment the Rescorla?Wagner model with an input representation learned by an
autoencoder. Since autoencoders perform probabilistic density modeling, this is probably
the most statistically minded of prior approaches to configural representation and has clear
parallels with our work.
Acknowledgments
This work was supported by National Science Foundation grants IIS-9978403 and DGE9987588. ND is funded by a Royal Society USA Research Fellowship and the Gatsby
Foundation. We thank Peter Dayan, Yael Niv and Geoff Gordon for helpful discussions.
References
[1] P. Dayan, T. Long, Advances in Neural Information Processing Systems 10 (1998), pp. 117?123.
[2] R. A. Rescorla, A. R. Wagner, Classical Conditioning II, A. H. Black, W. F. Prokasy, eds.
(Appleton-Century-Crofts, 1972), pp. 64?99.
[3] R. A. Rescorla, Journal of Comparative and Physiological Psychology 79, 307 (1972).
[4] J. M. Pearce, Psychological Review 101, 587 (1994).
[5] A. C. Courville, N. D. Daw, G. J. Gordon, D. S. Touretzky, Advances in Neural Information
Processing Systems 16 (2004).
[6] R. A. Rescorla, Quarterly Journal of Experimental Psychology 56B, 161 (2003).
[7] R. A. Rescorla, Animal Learning and Behavior 25, 200 (1997).
[8] E. S. Redhead, J. M. Pearce, Quarterly Journal of Experimental Psychology 48B, 46 (1995).
[9] E. F. Kremer, Journal of Experimental Psychology: Animal Behavior Processes 4, 22 (1978).
[10] J. B. Tenenbaum, T. L. Griffiths, Behavioral and Brain Sciences 24, 629 (2001).
[11] R. C. O?Reilly, J. W. Rudy, Psychological Review 108, 311 (2001).
[12] M. A. Gluck, C. Myers, Hippocampus 3, 491 (1993).
| 2711 |@word trial:21 version:1 stronger:1 seems:3 nd:1 extinction:1 hippocampus:1 additively:2 simulation:4 gradual:1 carry:1 configuration:2 contains:2 bc:12 suppressing:1 ours:1 activation:6 yet:2 profusion:1 written:1 readily:1 vor:1 realistic:1 refuted:1 update:1 progressively:1 discrimination:8 alone:4 generative:6 fewer:1 v:1 device:2 tone:9 ith:2 provides:1 preference:2 simpler:2 predecessor:3 differential:1 become:2 pairing:3 sustained:1 acti:1 behavioral:4 notably:2 behavior:4 nor:1 brain:1 retard:1 aaronc:1 food:2 resolve:1 param:1 underlying:1 competes:1 what:5 kind:1 finding:2 configural:27 quantitative:1 reinterpret:1 growth:1 shed:2 exactly:2 uk:1 control:4 unit:22 converse:1 grant:1 influencing:1 understood:1 treat:2 encoding:1 approximately:2 might:3 plus:1 black:1 resembles:1 studied:1 limited:1 range:2 statistically:2 unique:1 acknowledgment:1 spontaneously:1 testing:1 practice:1 block:8 differs:1 scant:1 empirical:1 thought:1 matching:1 reilly:1 integrating:2 griffith:1 anticipation:1 cannot:2 judged:1 context:1 influence:1 seminal:1 map:4 center:1 attention:1 focused:1 simplicity:2 rule:7 his:1 century:1 traditionally:1 variation:1 analogous:1 updated:2 laplace:2 spontaneous:1 trigger:1 play:2 substrate:1 pa:1 element:19 approximated:1 particularly:1 predicts:10 observed:12 role:2 solved:1 capture:4 connected:2 apportioned:1 ordering:1 decrease:1 highest:1 complexity:1 dynamic:1 renormalized:1 controversy:1 trained:8 weakly:2 carrying:1 predictive:1 division:1 basis:3 geoff:1 represented:3 additivity:1 describe:4 london:1 activate:3 monte:1 aggregate:1 larger:1 plausible:1 otherwise:2 favor:2 abstractly:1 jointly:2 itself:1 delivered:2 seemingly:1 hoc:3 advantage:1 myers:2 reinforcer:9 net:1 ucl:1 propose:1 rescorla:20 product:1 remainder:1 poorly:1 realistically:1 intuitive:1 differentially:1 recipe:1 elemental:5 cluster:1 produce:4 comparative:1 incremental:1 help:1 ac:1 measured:1 received:1 c:1 predicted:4 indicate:1 differ:2 concentrate:1 posit:1 direction:1 human:1 explains:2 generalization:14 niv:1 probable:1 summation:14 exploring:1 exp:1 great:1 cognition:1 predict:5 scope:1 claim:1 major:1 early:1 ambivalent:1 footshock:1 individually:2 minded:1 weighted:1 reflects:1 rather:4 arose:1 varying:1 broader:1 conjunction:2 encode:1 derived:1 exemplifies:1 bernoulli:1 likelihood:3 tradition:1 detect:1 posteriori:1 inference:13 helpful:1 dayan:2 typically:1 unlikely:1 explanatory:1 initially:2 hidden:1 reproduce:3 issue:10 classification:1 fidelity:2 overall:1 denoted:1 priori:3 augment:1 development:1 animal:21 constrained:1 integration:1 marginal:2 equal:1 sampling:1 progressive:1 future:3 mimic:1 stimulus:53 gordon:2 manipulated:1 simultaneously:2 national:1 individual:7 parsimoniously:1 phase:1 attempt:1 ab:18 certainly:1 dge9987588:1 admitting:1 light:11 chain:1 accurate:1 edge:1 partial:1 experience:7 guidance:2 weighs:1 theoretical:4 uncertain:1 psychological:2 instance:3 formalism:1 modeling:3 assertion:1 overshadowing:10 introducing:1 reflexive:1 neutral:1 rare:1 latents:3 successful:1 connect:1 answer:1 rudy:1 density:2 discriminating:1 probabilistic:2 participates:1 connecting:1 together:6 analogously:2 concrete:1 receipt:1 style:1 li:2 account:9 converted:1 includes:2 caused:1 ad:6 later:1 wm:10 bayes:2 elicited:2 parallel:1 ass:1 nathaniel:1 xor:2 reinforced:7 clarifies:1 identify:1 judgment:1 generalize:1 bayesian:6 accurately:1 produced:3 iid:1 carlo:1 drive:1 history:1 explain:4 simultaneous:1 strongest:1 touretzky:1 ed:1 against:2 nonetheless:1 acquisition:1 pp:2 associated:1 proved:1 treatment:2 improves:1 follow:2 response:9 unreinforced:2 though:2 strongly:2 evaluated:1 correlation:1 autoencoders:1 freezing:1 invite:1 nonlinear:1 reversible:1 readiness:1 lack:1 overlapping:2 perhaps:1 reveal:1 usa:1 effect:5 counterpart:2 illustrated:1 width:1 razor:1 hippocampal:2 stress:1 complete:1 demonstrate:1 interrelationship:1 consideration:2 novel:1 recently:1 sigmoid:2 common:1 conditioning:16 association:1 organism:1 discussed:1 occurred:1 accumulate:2 mellon:1 significant:1 appleton:1 framed:1 automatic:1 similarly:1 inclusion:1 phenomenological:1 dot:1 funded:1 similarity:5 impressive:1 longer:1 etc:3 depiction:1 posterior:3 recent:1 perspective:2 moderate:2 compound:37 binary:2 seen:1 additional:1 greater:1 determine:2 ii:2 full:3 sound:1 multiple:3 infer:1 match:2 faster:1 offer:1 long:1 concerning:2 ensuring:1 prediction:11 basic:1 circumstance:1 cmu:1 expectation:1 grounded:1 normalization:1 represent:2 robotics:1 achieved:1 fellowship:1 separately:1 addressed:2 wealth:1 grow:1 crucial:1 dwm:1 extra:2 rest:2 unlike:2 probably:1 subject:2 tend:1 incorporates:1 seem:3 leverage:1 rendering:1 affect:1 isolation:1 psychology:4 triggering:1 opposite:2 attenuated:2 tradeoff:2 absent:1 shift:1 inactive:1 whether:5 handled:1 akin:2 peter:1 cause:9 repeatedly:1 useful:1 detailed:1 clear:3 proportionally:2 tenenbaum:1 category:1 exist:1 neuroscience:1 estimated:1 correctly:2 materialize:1 broadly:2 carnegie:1 key:5 nevertheless:1 demonstrating:1 capital:1 changing:1 neither:1 ani:1 shock:4 contentious:1 sum:1 letter:1 uncertainty:4 dst:1 arrive:1 laid:1 almost:1 draw:1 delivery:1 prefer:1 capturing:1 uncontrolled:1 distinguish:1 courville:2 strength:4 occur:3 precisely:1 encodes:2 aspect:3 attempting:1 department:1 according:2 combination:1 remain:1 increasingly:1 unity:1 explained:2 equation:4 mutually:1 discus:2 turn:1 mechanistic:1 informal:1 studying:1 yael:1 probe:7 quarterly:2 away:1 differentiated:1 distinguished:2 original:1 responding:12 assumes:1 denotes:1 include:2 graphical:1 graded:1 approximating:1 classical:5 society:1 move:1 intend:1 question:3 added:5 quantity:2 fa:1 traditional:1 link:1 separate:3 thank:1 nx:3 considers:1 extent:2 reason:2 toward:2 stim:8 experiencedependent:1 modeled:1 relationship:1 balance:1 rise:2 implementation:1 perform:1 allowing:1 pearce:15 markov:1 behave:2 incorrectly:1 interacting:1 arbitrary:3 inferred:1 david:1 pair:3 required:1 cast:1 connection:2 learned:1 daw:2 address:1 beyond:2 suggested:1 below:1 pattern:17 mismatch:1 sparsity:1 including:1 royal:1 belief:3 unrealistic:1 overlap:8 event:6 greatest:1 difficulty:1 predicting:3 solvable:1 representing:1 scheme:1 historically:1 autoencoder:1 review:4 understanding:1 prior:10 geometric:1 determining:1 relative:2 lacking:1 prototypical:1 proportional:2 versus:1 foundation:6 integrate:1 degree:1 consistent:3 occam:1 cd:7 balancing:2 course:1 summary:1 accounted:1 supported:1 kremer:1 free:1 jth:1 formal:3 bias:4 understand:1 institute:1 explaining:1 wagner:15 distributed:1 curve:3 world:1 sensory:2 answerable:1 reinforcement:17 jump:1 founded:1 prokasy:1 sj:4 approximate:1 preferred:1 active:5 correlating:3 pittsburgh:1 assumed:2 discriminative:2 xi:4 spectrum:1 latent:25 why:3 promising:1 learn:6 transfer:8 nature:1 contributes:1 complex:2 reconciled:1 spread:1 linearly:1 whole:1 x1:4 augmented:3 body:1 wbias:2 gatsby:3 probing:1 inferring:1 candidate:1 croft:1 showing:1 normative:3 constellation:4 physiological:1 concern:2 weakest:1 incorporating:1 magnitude:1 conditioned:3 gluck:2 depicted:1 likely:3 neurophysiological:2 failed:1 expressed:1 saturating:1 tracking:1 collectively:1 chance:3 abc:12 conditional:1 viewed:1 presentation:7 goal:1 formulated:1 shared:2 change:1 experimentally:1 included:1 determined:3 averaging:2 justify:1 discriminate:4 experimental:11 aaron:1 college:1 puzzling:1 support:1 latter:1 arises:1 meant:1 assessed:2 mcmc:4 tested:2 phenomenon:6 correlated:3 |
1,885 | 2,712 | A harmonic excitation state-space approach to
blind separation of speech
Rasmus Kongsgaard Olsson and Lars Kai Hansen
Informatics and Mathematical Modelling
Technical University of Denmark, 2800 Lyngby, Denmark
rko,[email protected]
Abstract
We discuss an identification framework for noisy speech mixtures. A
block-based generative model is formulated that explicitly incorporates
the time-varying harmonic plus noise (H+N) model for a number of latent
sources observed through noisy convolutive mixtures. All parameters
including the pitches of the source signals, the amplitudes and phases of
the sources, the mixing filters and the noise statistics are estimated by
maximum likelihood, using an EM-algorithm. Exact averaging over the
hidden sources is obtained using the Kalman smoother. We show that
pitch estimation and source separation can be performed simultaneously.
The pitch estimates are compared to laryngograph (EGG) measurements.
Artificial and real room mixtures are used to demonstrate the viability
of the approach. Intelligible speech signals are re-synthesized from the
estimated H+N models.
1 Introduction
Our aim is to understand the properties of mixtures of speech signals within a generative
statistical framework. We consider convolutive mixtures, i.e.,
xt =
L?1
X
Ak st?k + nt ,
(1)
k=0
where the elements of the source signal vector, st , i.e., the ds statistically independent
source signals, are convolved with the corresponding elements of the filter matrix, Ak .
The multichannel sensor signal, xt , is furthermore degraded by additive Gaussian white
noise.
It is well-known that separation of the source signals based on second order statistics is
infeasible in general. Consider the second order statistic
hxt x>
t0 i =
L?1
X
>
Ak hst?k s>
t0 ?k0 iAk0 + R,
(2)
k,k0 =0
where R is the (diagonal) noise covariance matrix. If the sources can be assumed stationary
white noise, the source covariance matrix can be assumed proportional to the unit matrix
without loss of generality, and we see that the statistic is symmetric to a common rotation
of all mixing matrices Ak ? Ak U. This rotational invariance means that the acquired
statistic is not informative enough to identify the mixing matrix, hence, the source time
series.
However, if we consider stationary sources with known, non-trivial, autocorrelations
0
hst s>
t0 i = G(t ? t ), and we are given access to measurements involving multiple val0
ues of G(t ? t ), the rotational degrees of freedom are constrained and we will be able to
recover the mixing matrices up to a choice of sign and scale of each source time series.
Extending this argument by the observation that the mixing model (1) is invariant to filtering of a given column of the convolutive filter provided that the inverse filter is applied to
corresponding source signal, we see that it is infeasible to identify the mixing matrices if
these arbitrary inverse filters can be chosen to that they are allowed to ?whiten? the sources,
see also [1].
For non-stationary sources, on the other hand, the autocorrelation functions vary through
time and it is not possible to choose a single common whitening filter for each source. This
means that the mixing matrices may be identifiable from multiple estimates of the second
order correlation statistic (2) for non-stationary sources. Analysis in terms of the number
of free parameters vs. the number of linear conditions is provided in [1] and [2].
Also in [2], the constraining effect of source non-stationarity was exploited by the simultaneous diagonalization of multiple estimates of the source power spectrum. In [3] we
formulated a generative probabilistic model of this process and proved that it could estimate sources and mixing matrices in noisy mixtures. Blind source separation based on
state-space models has been studied, e.g., in [4] and [5]. The approach is especially useful
for including prior knowledge about the source signals and for handling noisy mixtures.
One example of considerable practical importance is the case of speech mixtures.
For speech mixtures the generative model based on white noise excitation may be improved
using more realistic priors. Speech models based on sinusoidal excitation have been quite
popular in speech modelling since [6]. This approach assumes that the speech signal is
a time-varying mixture of a harmonic signal and a noise signal (H+N model). A recent
application of this model for pitch estimation can be found in [7]. Also [8] and [9] exploit
the harmonic structure of certain classes of signals for enhancement purposes. A related
application is the BSS algorithm of [10], which uses the cross-correlation of the amplitude
in different frequency. The state-space model naturally leads to maximum-likelihood estimation using the EM-algorithm, e.g. [11], [12]. The EM algorithm has been used in related
models: [13] and [14].
In this work we generalize our previous work on state space models for blind source separation to include harmonic excitation and demonstrate that it is possible to perform simultaneous un-mixing and pitch tracking.
2 The model
The assumption of time variant source statistics help identify parameters that would otherwise not be unique within the model. In the following, the measured signals are segmented
into frames, in which they are assumed stationary. The mixing filters and observation noise
covariance matrix are assumed stationary across all frames.
The colored noise (AR) process that was used in [3] to model the sources is augmented to
include a periodic excitation signal that is also time-varying. The specific choice of periodic
basis function, i.e. the sinusoid, is motivated by the fact that the phase is linearizable,
facilitating one-step optimization. In frame n, source i is represented by:
sni,t
=
p
X
n n
fi,t
0 si,t?t0 +
t0 =1
=
p
X
K
X
n
n
n
?i,k
sin(?0,i
kt + ?in ) + vi,t
k=1
n n
fi,t
0 si,t?t0 +
t0 =1
K
X
n
n
n
cni,2k?1 sin(?0,i
kt) + cni,2k cos(?0,i
kt) + vi,t
(3)
k=1
n
where n ? {1, 2, .., N } and i ? {1, 2, .., ds }. The innovation noise, vi,t
, is i.i.d Gausn
sian. Clearly, (3) represents a H+N model. The fundamental frequency, ?0,i
, enters the
estimation problem in an inherent non-linear manner.
In order to benefit from well-established estimation theory, the above recursion is fitted into the framework of Gaussian linear models, see [15]. The Kalman filter model
is an instance of this model. The augmented state space is constructed by including a history of past samples for each source. Source vector i in frame n is defined:
?
?>
sni,t = sni,t sni,t?1 . . . sni,t?p+1
. All sni,t ?s are stacked in the total source vec?
?>
tor: ?snt = (sn1,t )> (sn2,t )> . . . (snds ,t )>
. The resulting state-space model is:
?snt
xnt
? tn
= Fn?snt?1 + Cn unt + v
n
n
= A?st + nt
? t ? N (0, Q), nt ? N (0, R) and ?sn1 ? N (?n , ?n ). The combined harmonics inwhere v
?>
?
, where the harmonics
put vector is defined: unt = (un1,t )> (un2,t )> . . . (unds ,t )>
corresponding to source i in frame n are:
?
?>
n
n
n
n
t) cos(?0,i
t) . . . sin(K?0,i
t) cos(K?0,i
t)
uni,t = sin(?0,i
It is apparent that the matrix multiplication by A
sources, where the dx ? ds channel filters are:
? >
..
a>
a11
12
>
? a>
..
a
22
? 21
A=? .
..
..
.
? .
.
.
a>
dx 1
In order to implement the
follows:
? n
???
F1 0
0
Fn2 ? ? ?
?
Fn = ?
..
..
? ...
.
.
0
0
???
? n
Q1 0
???
0
Qn2 ? ? ?
?
Qn = ?
..
..
? ...
.
.
0
0
???
? n
C1 0
???
0
Cn2 ? ? ?
?
Cn = ?
..
..
? ...
.
.
0
0
???
a>
dx 2
constitutes a convolutive mixing of the
?
a>
1ds
a>
2ds
..
.
?
?
?
?
a>
dx ds
..
H+N source model, the parameter matrices are constrained as
?
0
0
?
? ,
..
?
.
n
Fds
?
0
0
?
? ,
..
?
.
n
Qds
?
0
0
?
? ,
..
?
.
Cnds
?
?
?
Fni = ?
?
?
n
fi,1
1
0
..
.
n
fi,2
0
1
..
.
???
???
???
..
.
n
fi,p?1
0
0
..
.
n
fi,p
0
0
..
.
0
0
???
1
0
(Qni )jj 0 =
?
Cni
?
?
=?
?
?
n
qin
0
j = j 0W= 1
j 6= 1 j 0 6= 1
cni,1
0
0
..
.
cni,2
0
0
..
.
???
???
???
..
.
cni,2K
0
0
..
.
0
0
???
0
?
?
?
?
?
?
?
?
?
?
?
?
3 Learning
Having described the convolutive mixing problem in the general framework of linear Gaussian models, more specifically the Kalman filter model, optimal inference of the sources is
obtained by the Kalman smoother. However, since the problem at hand is effectively blind,
we also need to estimate the parameters. Along the lines of, e.g. [15], we will invoke an EM
approach. The log-likelihood is bounded
from below: L(?) ? F (?, p?) ?
p),
R
R J (?, p?) ? R(?
with the definitions J (?, p?) ? dS?
p(S) log p(X, S|?) and R(?
p) ? dS?
p(S) log p?(S).
In accordance with standard EM theory, J (?, p?) is optimized wrt. ? in the M-step. The
E-step infers the relevant moments of the marginal posterior, p? = p(S|X, ?). For the Gaussian model the means are also source MAP estimates. The combined E and M steps are
guaranteed not to decrease L(?).
3.1 E-step
The forward-backward recursions which comprise the Kalman smoother are employed in
the E-step to infer moments of the source posterior, p(S|X, ?), i.e. the joint posterior of
the sources conditioned on all observations. The relevant second-order statistic of this
n
distribution in segment n is the marginal posterior mean, ??st ? h?snt i, and autocorrelation,
n
n
n
>
n
n
n >
Mi,t ? hsi,t (si,t ) i ? [ mi,1,t mi,2,t .. mi,L,t ] , along with the marginal lag-one
1,n
>
>
n
n
covariance, M1,n
m1,n
.. m1,n
i,2,t
i,t ? hsi,t (si,t?1 ) i ? [ mi,1,t
i,L,t ] . In particular,
n
n
mi,t is the first element of mi,1,t . All averages are performed over p(S|X, ?). The forward
recursion also yields the log-likelihood, L(?).
3.2 M-step
The M-step utility function, J (?, p?), is defined:
J (?, p?) = ?
ds
ds
N
X
1XX
[
log det ?ni + (? ? 1)
log qin
2 n=1 i=1
i=1
+? log det R +
ds
X
h(sni,1 ? ?ni )T (?ni )?1 (sni,1 ? ?ni )i
i=1
+
ds
? X
X
t=2 i=1
h
?
X
1 n
n T n 2
(s
?
(d
)
z
)
i
+
h(xnt ? A?snt )T R?1 (xnt ? A?snt )i]
i
i,t
qin i,t
t=1
where h?i signifies averaging over the source posterior from the previous E-step, p(S|X, ?)
and ? is the frame length. The linear source parameters are grouped as
?>
?
?>
?
dni ? (fin )> (cni )>
, zni ? (sni,t?1 )> (uni,t )>
where
fin ? [ fi,1
fi,2
..
fi,p ]
>
,
cni ? [ ci,1
ci,2
>
.. ci,p ]
n
Optimization of J (?, p?) wrt. ? is straightforward (except for the ?0,i
?s). Relatively minor
changes are introduced to the estimators of e.g. [12] in order to respect the special constrained format of the parameter matrices and to allow for an external input to the model.
More details on the estimators for the correlated source model are given in [3].
n
It is in general difficult to maximize J (?, p?) wrt. to ?i,0
, since several local maxima exist,
n
e.g. at multiples of ?i,0
, see e.g. [6]. This problem is addressed by narrowing the search
range based on prior knowledge of the domain, e.g. that the pitch of speech lies in the range
n
50-400Hz. A candidate estimate for ?i,0
is obtained by computing the autocorrelation
n
n > n
function of si,t ? (fi ) si,t?1 . Grid search is performed in the vicinity of the candidate.
For each point in the grid we optimize dni :
" ? ?
?#?1 X
?
? ?
n
n >
X (Mn )
?
s
(u
)
mni,t,t?1
n
i,t?1
i,t?1
i,t
(4)
di,new =
s?ni,t uni,t
uni,t (?sni,t?1 )> uni,t (uni,t )>
t=2
t=2
F [Hz] (source 1)
At each step of the EM-algorithm, the parameters are normalized by enforcing ||Ai || = 1,
4000
3000
2000
1000
F [Hz] (source 2)
0
4000
3000
2000
1000
0
0
0.5
1
t [sec]
0
0.5
1
t [sec]
0
0.5
1
t [sec]
Figure 1: Amplitude spectrograms of the frequency range 0-4000Hz, from left to right: the
true sources, the estimated sources and the re-synthesized source.
that is enforcing a unity norm on the filter coefficients related to source i.
4 Experiment I: BSS and pitch tracking in a noisy artificial mixture
The performance of a pitch detector can be evaluated using electro-laryngograph (EGG)
recordings, which are obtained from electrodes placed on the neck, see [7]. In the following
experiment, speech signals from the TIMIT [16] corpus is used for which the EGG signals
were measured, kindly provided by the ?festvox? project (http://festvox.org).
Two male speech signals (Fs = 16kHz) were mixed through known mixing filters and
degraded by additive white noise (SNR ?20dB), constructing two observation signals. The
pitches of the speech signals were overlapping. The filter coefficients (of 2 ? 2 = 4 FIR
filter impulse responses) were:
?
?
1.00 0.35 ?0.20
0.00 0.00, 0.00 0.00 ?0.50 ?0.30 0.20
A =
0.00 0.00
0.70 ?0.20 0.15, 1.30 0.60
0.30
0.00 0.00
The signals were segmented into frames, ? = 320 ? 20ms, and the order of the ARprocess was set to p = 1. The number of harmonics was limited to K = 40. The pitch
grid search involved 30 re-estimations of dni . In figure 1 is shown the spectrograms of
F [Hz] (source 1)
160
140
0
120
100
160
140
120
100
0
F [Hz] (source 2)
80
80
0.4
0.6
0.8
t [sec]
1
Figure 2: The estimated (dashed) and EGG-provided (solid) pitches as a function of time.
The speech mixtures were artificially mixed from TIMIT utterances and white noise was
added.
approximately 1 second of 1) the original sources, 2) the MAP source estimates and 3) the
resynthesized sources (from the estimated model parameters). It is seen that the sources
were well separated. Also, the re-synthesizations are almost indistinguishable from the
source estimates. In figure 2, the estimated pitch of both speech signals are shown along
with the pitch of the EGG measurements.1 The voiced sections of the speech were manually
preselected, this step is easily automated. The estimated pitches do follow the ?true? pitches
as provided by the EGG. The smoothness of the estimates is further indicating the viability
of the approach, as the pitch estimates are frame-local.
5 Experiment II: BSS and pitch tracking in a real mixture
The algorithm was further evaluated on real room recordings that were also used in [17].2
Two male speakers synchronously count in English and Spanish (Fs = 16kHz). The mixtures were degraded with noise (SNR ?20dB). The filter length, the frame length, the order
of the AR-process and the number of harmonics were set to L = 25, ? = 320, p = 1 and
K = 40, respectively. Figure 3 shows the MAP source estimates and the re-synthesized
sources. Features of speech such as amplitude modulation are clearly evident in estimates
and re-synthesizations.3 A listening test confirms: 1) the separation of the sources and
2) the good quality of the synthesized sources, reconfirming the applicability of the H+N
model. Figure 4 displays the estimated pitches of the sources, where the voiced sections
were manually preselected. Although, the ?true? pitch is unavailable in this experiment, the
smoothness of the frame-local pitch-estimates is further support for the approach.
1
The EGG data are themselves noisy measurements of the hypothesized ?truth?. Bandpass filtering
was used for preprocessing.
2
The mixtures were obtained from http://inc2.ucsd.edu/?tewon/ica_cnl.html.
3
Note that the ?English? counter lowers the pitch throughout the sentence.
F [Hz] (source 1)
3000
2000
1000
F [Hz] (source 2)
0
3000
2000
1000
0
0
1
2
30
t [sec]
1
2
3
t [sec]
Figure 3: Spectrograms of the estimated (left) and re-synthesized sources (right) extracted
from the ?one two . . . ? and ?uno dos . . . ? mixtures, source 1 and 2, respectively
6 Conclusion
It was shown that prior knowledge on speech signals and quasi-periodic signals in general
can be integrated into a linear non-stationary state-space model. As a result, the simultaneous separation of the speech sources and estimation of their pitches could be achieved. It
was demonstrated that the method could cope with noisy artificially mixed signals and real
room mixtures. Future research concerns more realistic mixtures in terms of reverberation
time and inclusion of further domain knowledge. It should be noted that the approach is
computationally intensive, we are also investigating means for approximate inference and
parameter estimation that would allow real time implementation.
Acknowledgement
This work is supported by the Danish ?Oticon Fonden?.
References
[1] E. Weinstein, M. Feder and A.V. Oppenheim, Multi-channel signal separation by decorrelation,
IEEE Trans. on speech and audio processing, vol. 1, no. 4, pp. 405-413,1993.
[2] Parra, L., Spence C., Convolutive blind separation of non-stationary sources. IEEE Trans. on
speech and audio processing, vol. 5, pp. 320-327, 2000.
[3] Olsson, R. K., Hansen L. K., Probabilistic blind deconvolution of non-stationary source. Proc.
EUSIPCO, 2004, accepted. Olsson R. K., Hansen L. K., Estimating the number of sources in a
noisy convolutive mixture using BIC. International conference on independent component analysis 2004, accepted. Preprints may be obtained from http://www.imm.dtu.dk/?rko/
research.htm.
[4] Gharbi, A.B.A., Salam, F., Blind separtion of independent sources in linear dynamical media. NOLTA, Hawaii, 1993. http://www.egr.msu.edu/bsr/papers/blind_
separation/nolta93.pdf
120
100
0
F [Hz] (source 2)
0
F [Hz] (source 1)
140
80
180
160
140
120
100
80
0
2
4
t [sec]
6
8
Figure 4: Pitch tracking in ?one two . . . ?/?uno dos . . . ? mixtures.
[5] Zhang, L., Cichocki, A., Blind Deconvolution of dynamical systems: a state space appraoch,
Journal of signal processing, vol. 4, no. 2, pp. 111-130, 2000.
[6] McAulay, R.J., Quateri. T.F., Speech analysis/synthesis based on a sinusoidal representation,
IEEE Trans. on acoustics, speech and signal processing, vol. 34, no. 4, pp. 744-754, 1986.
[7] Parra, L., Jain U., Approximate Kalman filtering for the harmonic plus noise model. IEEE Workshop on applications of signal processing to audio and acoustics, pp. 75-78, 2001.
[8] Nakatani, T., Miyoshi, M., and Kinoshita, K., One microphone blind dereverberation based on
quasi-periodicity of speech signals, Advances in Neural Information Processing Systems 16 (to
appear), MIT Press, 2004.
[9] Hu, G. Wang, D., Monaural speech segregation based on pitch tracking and amplitude modulation, IEEE Trans. neural networks, in press, 2004.
[10] Anem?uller, J., Kollmeier, B., Convolutive blind source separation of speech signals based on
amplitude modulation decorrelation, Journal of the Acoustical Society of America, vol. 108, pp.
2630, 2000.
[11] A. P. Dempster, N. M. Laird, and Rubin D. B., Maximum liklihood from incomplete data via
the EM algorithm, Journal of the Royal Statistical Society, vol. 39, pp. 1?38, 1977.
[12] Shumway, R.H., Stoffer, D.S., An approach to time series smoothing and forecasting using the
EM algorithm. Journal of time series analysis, vol. 3, pp. 253-264. 1982.
[13] Moulines E., Cardoso J. F., Gassiat E., Maximum likelihood for blind separation and deconvolution of noisy signals using mixture models, ICASSP, vol. 5, pp. 3617-20, 1997.
[14] Cardoso, J.F., Snoussi, H. , Delabrouille, J., Patanchon, G., Blind separation of noisy Gaussian
stationary sources. Application to cosmic microwave background imaging, Proc. EUSIPCO, pp
561-564, 2002.
[15] Roweis, S., Ghahramani, Z., A unifying review of linear Gaussian models. Neural Computation,
vol. 11, pp. 305-345, 1999.
[16] Center for Speech Technology Research, University of Edinburgh,http://www.cstr.ed.
ac.uk/
[17] Lee, T.-W., Bell, A.J., Orglmeister, R., Blind source separation of real world signals, Proc. IEEE
international conference neural networks, pp 2129-2135, 1997.
| 2712 |@word norm:1 hu:1 confirms:1 covariance:4 q1:1 solid:1 moment:2 series:4 past:1 nt:3 si:6 dx:4 fn:2 realistic:2 additive:2 informative:1 v:1 stationary:10 generative:4 fni:1 colored:1 org:1 zhang:1 mathematical:1 along:3 constructed:1 weinstein:1 autocorrelation:3 manner:1 acquired:1 themselves:1 multi:1 moulines:1 provided:5 xx:1 bounded:1 project:1 estimating:1 medium:1 sni:10 nakatani:1 uk:1 unit:1 appear:1 accordance:1 local:3 eusipco:2 ak:5 modulation:3 approximately:1 plus:2 studied:1 co:3 limited:1 range:3 statistically:1 practical:1 unique:1 spence:1 block:1 implement:1 bell:1 put:1 preprints:1 www:3 optimize:1 map:3 fn2:1 demonstrated:1 center:1 straightforward:1 estimator:2 exact:1 us:1 element:3 observed:1 narrowing:1 enters:1 snt:6 wang:1 decrease:1 linearizable:1 counter:1 dempster:1 unt:2 segment:1 basis:1 easily:1 joint:1 htm:1 k0:2 icassp:1 represented:1 america:1 stacked:1 separated:1 jain:1 artificial:2 quite:1 apparent:1 kai:1 lag:1 otherwise:1 statistic:8 noisy:10 laird:1 snds:1 un1:1 qin:3 relevant:2 mixing:13 roweis:1 qni:1 enhancement:1 electrode:1 extending:1 a11:1 help:1 miyoshi:1 ac:1 measured:2 minor:1 hst:2 rko:2 filter:15 lars:1 cstr:1 f1:1 parra:2 appraoch:1 tor:1 vary:1 purpose:1 estimation:8 proc:3 hansen:3 grouped:1 uller:1 mit:1 clearly:2 sensor:1 gaussian:6 aim:1 varying:3 lkh:1 modelling:2 likelihood:5 kollmeier:1 inference:2 integrated:1 hidden:1 quasi:2 html:1 constrained:3 special:1 smoothing:1 marginal:3 comprise:1 having:1 manually:2 represents:1 constitutes:1 future:1 inherent:1 simultaneously:1 olsson:3 phase:2 freedom:1 stationarity:1 stoffer:1 male:2 mixture:21 kt:3 microwave:1 incomplete:1 re:7 fitted:1 instance:1 column:1 ar:2 signifies:1 applicability:1 snr:2 periodic:3 combined:2 st:4 fundamental:1 international:2 probabilistic:2 invoke:1 informatics:1 lee:1 synthesis:1 choose:1 fir:1 hawaii:1 external:1 sinusoidal:2 sec:7 coefficient:2 explicitly:1 blind:13 vi:3 performed:3 recover:1 laryngograph:2 voiced:2 timit:2 ni:5 degraded:3 yield:1 identify:3 cni:8 generalize:1 identification:1 sn1:2 history:1 simultaneous:3 detector:1 oppenheim:1 ed:1 danish:1 definition:1 frequency:3 involved:1 pp:12 naturally:1 mi:7 di:1 proved:1 popular:1 knowledge:4 infers:1 amplitude:6 follow:1 response:1 improved:1 evaluated:2 generality:1 furthermore:1 correlation:2 d:12 hand:2 overlapping:1 autocorrelations:1 quality:1 impulse:1 effect:1 hypothesized:1 normalized:1 true:3 gharbi:1 hence:1 vicinity:1 sinusoid:1 symmetric:1 white:5 sin:4 indistinguishable:1 spanish:1 speaker:1 excitation:5 whiten:1 noted:1 m:1 pdf:1 evident:1 demonstrate:2 tn:1 harmonic:10 fi:10 common:2 rotation:1 khz:2 m1:3 synthesized:5 measurement:4 vec:1 ai:1 smoothness:2 grid:3 inclusion:1 access:1 whitening:1 posterior:5 recent:1 certain:1 exploited:1 seen:1 spectrogram:3 employed:1 egr:1 maximize:1 dashed:1 signal:33 smoother:3 multiple:4 hsi:2 ii:1 infer:1 segmented:2 technical:1 cross:1 pitch:24 involving:1 variant:1 achieved:1 cosmic:1 c1:1 background:1 addressed:1 source:72 hz:10 recording:2 resynthesized:1 electro:1 db:2 incorporates:1 constraining:1 viability:2 enough:1 automated:1 bic:1 cn:2 intensive:1 det:2 listening:1 t0:7 motivated:1 utility:1 feder:1 forecasting:1 f:2 speech:27 jj:1 useful:1 cardoso:2 sn2:1 multichannel:1 http:5 exist:1 fds:1 sign:1 estimated:9 vol:9 backward:1 imaging:1 inverse:2 almost:1 throughout:1 separation:14 guaranteed:1 display:1 identifiable:1 mni:1 uno:2 argument:1 relatively:1 format:1 across:1 em:8 unity:1 kinoshita:1 invariant:1 handling:1 lyngby:1 computationally:1 segregation:1 discus:1 count:1 wrt:3 convolved:1 original:1 assumes:1 include:2 unifying:1 exploit:1 ghahramani:1 especially:1 society:2 added:1 diagonal:1 acoustical:1 trivial:1 enforcing:2 denmark:2 kalman:6 length:3 rasmus:1 rotational:2 innovation:1 difficult:1 reverberation:1 xnt:3 implementation:1 perform:1 zni:1 observation:4 fin:2 frame:10 synchronously:1 ucsd:1 monaural:1 arbitrary:1 salam:1 introduced:1 optimized:1 sentence:1 acoustic:2 established:1 trans:4 able:1 below:1 dynamical:2 convolutive:8 dereverberation:1 preselected:2 orglmeister:1 including:3 royal:1 power:1 decorrelation:2 sian:1 recursion:3 mn:1 tewon:1 technology:1 dtu:2 cichocki:1 ues:1 utterance:1 prior:4 review:1 acknowledgement:1 multiplication:1 shumway:1 loss:1 mixed:3 proportional:1 filtering:3 degree:1 cn2:1 nolta:1 rubin:1 periodicity:1 placed:1 supported:1 free:1 english:2 infeasible:2 allow:2 understand:1 dni:3 benefit:1 edinburgh:1 bs:3 world:1 qn:1 forward:2 preprocessing:1 cope:1 approximate:2 uni:6 imm:2 investigating:1 corpus:1 assumed:4 spectrum:1 un:1 latent:1 search:3 msu:1 channel:2 unavailable:1 artificially:2 constructing:1 domain:2 kindly:1 intelligible:1 noise:14 gassiat:1 allowed:1 facilitating:1 augmented:2 egg:7 bandpass:1 lie:1 candidate:2 xt:2 specific:1 dk:2 concern:1 deconvolution:3 workshop:1 hxt:1 effectively:1 importance:1 ci:3 diagonalization:1 conditioned:1 tracking:5 truth:1 extracted:1 formulated:2 room:3 considerable:1 change:1 specifically:1 except:1 averaging:2 microphone:1 total:1 neck:1 invariance:1 accepted:2 indicating:1 support:1 audio:3 correlated:1 |
1,886 | 2,713 | The Entire Regularization Path for the Support
Vector Machine
Trevor Hastie
Department of Statistics
Stanford University
Stanford, CA 94305, USA
[email protected]
Saharon Rosset
IBM Watson Research Center
P.O. Box 218
Yorktown Heights, N.Y. 10598
[email protected]
Robert Tibshirani
Department of Statistics
Stanford University
Stanford, CA 94305, USA
[email protected]
Ji Zhu
Department of Statistics
University of Michigan
Ann Arbor, MI 48109-1092
[email protected]
Abstract
In this paper we argue that the choice of the SVM cost parameter can be
critical. We then derive an algorithm that can fit the entire path of SVM
solutions for every value of the cost parameter, with essentially the same
computational cost as fitting one SVM model.
1
Introduction
We have a set of n training pairs xi , yi , where xi ? Rp is a p-vector of real valued predictors
(attributes) for the ith observation, yi ? {?1, +1} codes its binary response. The standard
criterion for fitting the linear SVM )[1, 2, 3] is
1
min ||?||2 + C
?i ,
?0 ,? 2
i=1
n
subject to, for each i: yi (?0 + xTi ?)
(1)
?
1 ? ?i .
Here the ?i are non-negative slack variables that allow points to be on the wrong side of their
?soft margin? (f (x) = ?1), as well as the decision boundary, and C is a cost parameter that
controls the amount of overlap. If the data are separable, then for sufficiently large C the
solution achieves the maximal margin separator; if not, the solution achieves the minimum
overlap solution with largest margin.
Alternatively, we can formulate the problem using a (hinge) Loss + Penalty criterion [4, 5]:
min
?0 ,?
n
i=1
[1 ? yi (?0 + ? T xi )]+ +
?
||?||2 .
2
The regularization parameter ? in (2) corresponds to 1/C, with C in (1).
(2)
This latter formulation emphasizes the role of regularization. In many situations we have
sufficient variables (e.g. gene expression arrays) to guarantee separation. We may nevertheless avoid the maximum margin separator (? ? 0), which is governed by observations
on the boundary, in favor of a more regularized solution involving more observations.
The nonlinear
kernel SVMs can be represented in this form as well. With kernel K and
n
f (x) = ?0 + i=1 ?i K(x, xi ), we solve [5]
min
?0 ,?
n
[1 ? yi (?0 +
i=1
n
?
?j ?j K(xj , xj ).
2 j=1
n
?i K(xi , xj ))] +
j=1
n
(3)
j =1
Often the regularization parameter C (or ?) is regarded as a genuine ?nuisance?. Software
packages, such as the widely used SVMlight [6], provide default settings for C.
To illustrate the effect of regularization, we generated data from a pair of mixture densities,
described in detail in [5, Chapter 2]. We used an SVM with a radial kernel K(x, x ) =
exp(??||x ? x ||2 ). Figure 1 shows the test error as a function of C for these data, using
four different values for ?. Here we see a dramatic range in the correct choice for C (or
? = 1/C). When ? = 5, the most regularized model is called for; when ? = 0.1, the least
regularized.
Test Error Curves ? SVM with Radial Kernel
?=1
? = 0.5
? = 0.1
0.30
0.25
0.20
Test Error
0.35
?=5
1e?01
1e+01
1e+03
1e?01
1e+01
1e+03
1e?01
1e+01
1e+03
1e?01
1e+01
1e+03
C = 1/?
Figure 1: Test error curves for the mixture example, using four different values for the radial kernel
parameter ?.
One of the reasons that investigators avoid extensive exploration of C is the computational
cost involved. In this paper we develop an algorithm which fits the entire path of SVM
solutions [?0 (C), ?(C)], for all possible values of C, with essentially the computational
cost of fitting a single model for a particular value of C. Our algorithm exploits the fact
that the Lagrange multipliers implicit in (1) are piecewise-linear in C. This also means that
?
the coefficients ?(C)
are also piecewise-linear in C. This is true for all SVM models, both
linear and nonlinear kernel-based SVMs.
2
Problem Setup
We use a criterion equivalent to (1), implementing the formulation in (2):
n
?
min
?i + ? T ? subject to 1 ? yi f (xi ) ? ?i ; ?i ? 0; f (x) = ?0 + ? T x.
?,?0
2
i=1
(4)
Initially we consider only linear SVMs to get the intuitive flavor of our procedure; we then
generalize to kernel SVMs.
We construct the Lagrange primal function
LP :
n
n
n
? T
?i + ? ? +
?i (1 ? yi f (xi ) ? ?i ) ?
?i ?i
2
i=1
i=1
i=1
(5)
and set the derivatives to zero. This gives
?
:
??
?
:
??0
1
?i yi xi
? i=1
n
?=
n
yi ?i = 0,
(6)
(7)
i=1
along with the KKT conditions
?i (1 ? yi f (xi ) ? ?i ) = 0
?i ?i = 0
1 ? ?i ? ?i = 0
(8)
(9)
(10)
We see that 0 ? ?i ? 1, with ?i = 1 when ?i > 0 (which is when yi f (xi ) < 1). Also
when yi f (xi ) > 1, ?i = 0 since no cost is incurred, and ?i = 0. When yi f (xi ) = 1, ?i
can lie between 0 and 1.
The usual Lagrange multipliers associated with the solution to (1) are ?i = ?i /? = C?i .
We prefer our formulation here since our ?i ? [0, 1], and this simplifies the definition of
the paths we define.
We wish to find the entire solution path for all values of ? ? 0. Our basic idea is as follows.
We start with ? large and decrease it toward zero, keeping track of all the events that occur
along the way. As ? decreases, ||?|| increases, and hence the width of the margin decreases.
As this width decreases, points move from being inside to outside their margins. Their
corresponding ?i change from ?i = 1 when they are inside their margin (yi f (xi ) < 1) to
?i = 0 when they are outside their margin (yi f (xi ) > 1). By continuity, points must linger
on the margin (yi f (xi ) = 1) while their ?i decrease from 1 to 0. We will see that the ?i (?)
trajectories are piecewise-linear in ?, which affords a great computational savings: as long
as we can establish the break points, all values in between can be found by simple linear
interpolation. Note that points can return to the margin, after having passed through it.
It is easy to show that if the ?i (?) are piecewise linear in ?, then both ?i (C) = C?i (C)
and ?(C) are piecewise linear in C. It turns out that ?0 (C) is also piecewise linear in C.
Our algorithm keeps track of the following sets:
? M = {i : yi f (xi ) = 1, 0 ? ?i ? 1}, M for Margin
? I = {i : yi f (xi ) < 1, ?i = 1}, I for Inside the margin
? O = {i : yi f (xi ) > 1, ?i = 0}, O for Outside the margin
3
The Algorithm
Due to space restrictions, we show some details here; the rest can be found in [7].
Initialization
The initial conditions depend on whether the classes are balanced or not (n+ = n? ). The
balanced case is easier. For very large ?, ||?|| is small, and the the margin is very wide,
all points are in O, and hence ?i = 1?i. From (6) this means the orientation of ? is fixed
until the ?i change. The margin narrows as ? decreases, but the orientation remains fixed.
Because of (7), the narrowing margin must connect with an outermost member of each
class simultaneously. These points are easily identified, and this establishes the first event,
the first tenants of M, and ?0 .
When n? = n+ , the setup is more complex. In order to satisfy the constraint (7), a
quadratic programming algorithm is needed to obtain the initial configuration. See [7] for
details.
Kernels
The development so far has been in the original feature space. It is easy to see that the
entire development carries through with ?kernels? as well. In this case f (x) = ?0 + g(x),
and the only change that occurs is that (6) is changed to
1
?j yj K(xi , xj ), i = 1, . . . , n,
? j=1
n
g(xi ) =
(11)
or ?j (?) = ?j yj /? using the notation in (3). Hereafter we will develop our algorithm for
this more general kernel case.
The Path
The algorithm hinges on the set of points M sitting on the margin. We consider M at the
point that an event has occurred:
1. The initial event, which means 2 or more points start in M, with their initial values
of ? ? [0, 1].
2. A point from I has just entered M, with its value of ?i initially 1.
3. A point from O has reentered M, with its value of ?i initially 0.
4. One or more points in M has left the set, to join either O or I.
Whichever the case, for continuity reasons this set will stay stable until the next event
occurs, since to pass through M, a point?s ?i must change from 0 to 1 or vice versa. Since
all points in M have yi f (xi ) = 1, we can establish a path for their ?i .
We use the subscript to index the sets above immediately after the th event has occurred.
Suppose |M | = m, and let ?i , ?0 and ? be the values of these parameters at the point of
entry. Likewise f is the function at this point. For convenience we define ?0 = ??0 , and
hence ?0 = ? ?0 .
Since
?
?
n
1 ?
yj ?j K(x, xj ) + ?0 ? ,
f (x) =
? j=1
for ? > ? > ?+1 we can write
?
?
f (x) =
f (x) ? f (x) + f (x)
?
?
?
?
1?
=
(?j ? ?j )yj K(x, xj ) + (?0 ? ?0 ) + ? f (x)? .
?
j?M
(12)
(13)
The second line follows because all the observations in I have their ?i = 1, and those in
O have their ?i = 0, for this range of ?. Since each of the m points xi ? M are to stay
on the margin, we have that
?
?
1?
(?j ? ?j )yi yj K(xi , xj ) + yi (?0 ? ?0 ) + ? ? = 1, ?i ? M .
(14)
?
j?M
Writing ?j = ?j ? ?j , from (14) we have
?j yi yj K(xi , xj ) + yi ?0 = ? ? ?, ?i ? M .
(15)
j?M
Furthermore, since at all times
n
i=1
yi ?i = 0, we have that
yj ?j = 0.
(16)
j?M
Equations (15) and (16) constitute m + 1 linear equations in m + 1 unknowns ?j , and can
be solved. The ?j and hence ?j will change linearly in ?, until the next event occurs:
?j = ?j ? (? ? ?)bj , j ? {0} ? M .
(17)
See [7] for more precise details on solving these equations.
From (13) we have
f (x) =
?
f (x) ? h (x) + h (x),
?
where
h (x) =
yj bj K(x, xj ) + b0
(18)
(19)
j?M
Thus the function itself changes in a piecewise-inverse manner in ?.
Finding ?+1
The paths continue until one of the following events occur:
1. One of the ?i for i ? M reaches a boundary (0 or 1). For each i the value of ?
for which this occurs is easily established.
2. One of the points in I or O attains yi f (xi ) = 1.
By examining these conditions, we can establish the largest ? < ? for which an event
occurs, and hence establish ?+1 and update the sets.
Termination
In the separable case, we terminate when I becomes empty. At this point, all the ?i in (4)
are zero, and further movement increases the norm of ? unnecessarily.
In the non-separable case, ? runs all the way down to zero. For this to happen without f
?blowing up? in (18), we musthave f ? h = 0, and hence the boundary and margins
remain fixed at a point where i ?i is as small as possible, and the margin is as wide as
possible subject to this constraint.
3.1
Computational Complexity
?=1
? = 0.5
? = 0.1
1e?04
1e?02
1e+00
?
1e?03
Eigenvalue
?=5
1e?07
1 1111111 11111
11 111
1
111
11111
11
11
1e?11
80
60
40
Margin Size
20
0
4
3
23
43
2
2
33
422
1111
222
33322
41111
332211111
4 1111
2 111111111111111111
33222
3333222222
44
11111111111111
1111111111111111
33 2222222
4
11111111111
22222
3333
11111111111
4
33
22222
11111111111
44
3333
22
1111111111
22222
333
4
111111111
222
4
11111111
333
2222
3333
11111111
22222
444
111111
333
2222
111111
4
333
2222
1111111
44
3333
222222
11111
3
11111
2222
44
33
11111
2222
44
33
11111
33
22222
4
33333
1111
22222
44
33
11
22
1111
333
444
2222
1111
2222
333
222
44
3333
1111
22
33
22
111
4
333
222
444
1
333
222
4
22
111
33
2
4
33333
4
2222
1
22
44
33
11
2222
33
44
333
2222
4
333
222
44
33
222
44
33
222
44
3333
222
33
22
44
333
222
44
2222
33
4
333
44
22
333
222
444
2
33
4
222
33
44
333
22
4
222
33
4
333
44
222
333
44
444
33
222
33
44
222
333
222
333
444
3
222
4
333
22
4444
2
3333
444
22
3
4
2
3333
44444444444444444444444444444444444444444444444444444444444444444444444444444444444444444444444444444444444444444
2
22
3
2
4
333333333333333333333333333333333333333333
222222222
4
3
2
?=5
1 1 1
11
11 11
1111
11 11 11111
111
1 111
111111 1 1
1111 11
11 1
111 1
111111111 1
1 11 1 111 1
11 1 1111
1111 1 1 11
1
11
1 1 1
1111 1
1 1
111111
111111 1
11111
1111
1
111
11111
11 111111
111
11
1111111
11
1
11111
11 11
1111
1
1
111
1111
1111
1111
11
1111
11
11
11
1
1
11
1111
1
1
1111
111
11
1
11
1
11
1
1
1
1
1
11
11
11111
1
1
2222
2222
11
111
2 2222 2
2
222 22
2
1
111
2 2 22 2 222 2
222
2222
1111
222222222
22222
2222 22222 222
2
2
2
2
2
2
2
2
2
2
2
11
11 111
2
2 222222
2 22
2
22 2
22222
2222
22
22 2 2222
1
111
111
1
1
111111
1
2222 22222222 22 2 2 22222 22
22
2
1
33
33
2
11 111111111
22
222222222 22
22 22
222
33 3
3333 3
222
22
1
22
22222222222222
22 2222
11
11111
3333 3 33
33333333333 3
3 333 33
3
33
1
22
22
111
3
3333333 33333
3333
2
2
2
2
2
2
22222
222
1
1
333
3
33 3 3
2222
22222222
22 22
22
22 2
22
22 2
222222
222
111
11
1
3
33333
2 2 2 2 22
2
2
2 2222 222
2222
222222
22 22 22222222
2222
22 22222
11
333 3333333
3
2
2
2
2
33333
33
333 33
222
22222 22
22
2 22 22 2222 22222
11
3 33
333 3333
333
2
3 33333
333333333
333
222 22222
2
22 2
222
222
11
3
33333333333 3
3333333 33
333 3
222
2
2
2222
2
2
33 3333 333
3
3
3
2
2
2
2
2
2
1
33
3
3
3
22
2 22
22
222222222
1
3333
333
33333
333
33333 33
33333
33
1
3333333
2
2222
22
11
3
333 33333
22
2
3
333 333333333
333
2
22 2
22222 22222
11
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
2
2
3 333333
33
3 333333333333 3
333333333
333 3 33333 3
22222222
222222
22
4 4 4
4
333 333
222
22211
44444 444444
33
3 333333
3 3
333333333 333 33
33
33
3
2
2 222
22
2
222
4 444 444
4444444444
1
3
3
33333
33333 3
3333333
33 22
2
2
2
222
22
4444 4 4444444444 4
4
4 44 444444
44
1
3
3
3
3
3
3
3
3
2
2
2
2
222
22
22
2
4
4
11
1212
1
33
3
2
2
4
4 44
44 44 4 4
11
1
333 3333 3 3
33 33
333
33 2
222 22
222
2
2
2
2
2
212
22 222
44
4 4444444 44 4 4
44
44
1
1
1
33
3333333333333
22
2
444 4
4
4
44
4
1
1
111
3
222
2222 22
44 44
4 4 44
44 44444444
444
4
4 4 44
4444 4444444
1
3
3 333
33
21 1
222
4
444
444
444
11
11
12
3
3
222
22
22
222
22
22
44 4 4 444
4444
4
4
4
4
4
4
4
4
4
4
4
4
4
3
333
33233
222
2
2
4 4
4444 444 44 444 4444
44444444 44
4 4
4 4 44
4444
1
3333333
3
3332
3
3
3
33
3222 33 3
2
23
22
23
233
44
13
1
31
333
32
3
33
3
2
2
2
4
4
444 4
444 444 444
44444 44
4 4 444 44
322
3
333
31
3
323
33333
3
3333
23
2
23
223
44
4 4 4 444 44 444444
444 33333333 333333
111
13
13
23333
3
3
3
34
3
444
44 4
4
4
4
4
4
4
4
4
3
3
334
3333444444441
33
32
32
34
2
232
24233
2
44
4 4
44
44
41
4
441434
43144444342
1
2
234
3
243
2
42
4
4
44 4444444 44444 44444444444
44
444
2
43
42
44
44
4444 4114414
41444114
444444443
1
2
44
1
11
1 1 11
1e?15
100
1e+01
At any update event along the path of our algorithm, the main computational burden is
solving the system of equations of size m = |M |. While this normally involves O(m3 )
computations, since M+1 differs from M by typically one observation, inverse updating can reduce the computations to O(m2 ). The computation of h (xi ) in (19) requires
O(nm ) computations. Beyond that, several checks of cost O(n) are needed to evaluate
the next move.
?=1
? = 0.5
? = 0.1
0
50
100
150
200
Sequence Number
Figure 2: [Left] The margin sizes |M | as a function of ?, for different values of the radial-kernel
parameter ?. The vertical lines show the positions used to compare the times with libsvm. [Right]
The eigenvalues (on the log scale) for the kernel matrices K? corresponding to the four values
of ?.The larger eigenvalues correspond in this case to smoother eigenfunctions, the small ones to
rougher. The rougher eigenfunctions get penalized exponentially more than the smoother ones. For
smaller values of ?, the effective dimension of the space is truncated.
Although we have no hard results, our experience so far suggests that the total number ?
of moves is O(k min(n+ , n? )), for k around 4 ? 6; hence typically some small multiple
c of n. If the average size of M is m, this suggests the total computational burden is
O(cn2 m + nm2 ), which is similar to that of a single SVM fit.
Our R function SvmPath computes all 632 steps in the mixture example (n+ = n? =
100, radial kernel, ? = 1) in 1.44(0.02) secs on a Pentium 4, 2Ghz Linux machine; the svm
function (using the optimized code libsvm, from the R library e1071) takes 9.28(0.06)
seconds to compute the solution at 10 points along the path. Hence it takes our procedure
about 50% more time to compute the entire path, than it costs libsvm to compute a typical
single solution.
4
Mixture simulation continued
The ? in Figure 1 are the entire collection of change points as described in Section 3. We
were at first surprised to discover that not all these sequences achieved zero training errors
on the 200 training data points, at their least regularized fit. In fact the minimal training errors, and the corresponding values for ? are summarized in Table 1. It is sometimes argued
that the implicit feature space is ?infinite dimensional? for this kernel, which suggests that
perfect separation is always possible. The last row of the table shows the effective rank of
the 200?200 kernel Gram matrix K (which we defined to be the number of singular values
greater than 10?12 ). In general a full rank K is required to achieve perfect separation. This
rank-deficiency of the Gram matrix has been noted by a number of other authors.
This emphasizes the fact that not all features in the feature map implied by K are of equal
?
Training Errors
Effective Rank
5
0
200
1
12
177
0.5
21
143
0.1
33
76
Table 1: The number of minimal training errors for different values of the radial kernel scale parameter ?, for the mixture simulation example. Also shown is the effective rank of the 200 ? 200 Gram
matrix K? .
stature; many of them are shrunk way down to zero. Rephrasing, the regularization in (3)
penalizes unit-norm features by the inverse of their eigenvalues, which effectively annihilates some, depending on ?. Small ? implies wide, flat kernels, and a suppression of
wiggly, ?rough? functions.
Writing (3) in matrix form,
min L[y, K?] +
?0 ,?
? T
? K?,
2
(20)
we reparametrize using the eigen-decomposition of K = UDUT . Let K? = U? ? where
? ? = DUT ?. Then (20) becomes
min? L[y, U? ? ] +
?0 ,?
? ? T ?1 ?
? D ? .
2
(21)
Now the columns of U are unit-norm basis functions (in R2 ) spanning the column space
of K; from (21) we see that those members corresponding to near-zero eigenvalues (the
elements of the diagonal matrix D) get heavily penalized and hence ignored. Figure 2
shows the elements of D for the four values of ?.
5
Discussion
Our work on the SVM path algorithm was inspired by early work on exact path algorithms
in other settings. ?Least Angle Regression? [8] show that the coefficient path for the sequence of ?lasso? coefficients is piecewise linear. The lasso uses a quadratic criterion,
with an L1 constraint. In fact, any model with an L1 constraint and a quadratic, piecewise
quadratic, piecewise linear, or mixed quadratic and linear loss function, will have piecewise
linear coefficient paths, which can be calculated exactly and efficiently for all values of ?
[9]. This includes the L1 SVM [10].
The SVM model has a quadratic constraint and a piecewise linear (?hinge?) loss function.
This leads to a piecewise linear path in the dual space, hence the Lagrange coefficients ?i
are piecewise linear.
Of course, quadratic criterion + quadratic constraints also lead to exact path solutions, as
in the classic ridge regression case, since a closed form solution is obtained via the SVD.
The general techniques employed in this paper are known as parametric programming in
convex optimization. After completing this work, it was brought to our attention that [11]
reported on the picewise-linear nature of the lagrange multipliers, although they did not
develop the path algorithm. [12, 13] employ techniques similar to ours in incremental
learning for SVMs. These authors do not construct exact paths as we do, but rather focus
on updating and downdating the solutions as more (or less) data arises. [14] allow for
updating the parameters as well, but again do not construct entire solution paths.
The SvmPath has been implemented in the R computing environment, and is available
from the R website.
Acknowledgements
The authors thank Jerome Friedman for helpful discussions, and Mee-Young Park for assisting with some of the computations. Trevor Hastie was partially supported by grant
DMS-0204162 from the National Science Foundation, and grant RO1-EB0011988-08 from
the National Institutes of Health.
References
[1] B. Boser, I. Guyon, and V. Vapnik. A training algorithm for optimal margin classifiers.
In Proceedings of COLT II, Philadelphia, PA, 1992.
[2] C. Cortes and V. Vapnik. Support vector networks. Machine Learning, 20:1?25, 1995.
[3] Bernard Sch?olkopf and Alex Smola. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond (Adaptive Computation and Machine Learning). MIT Press, 2001.
[4] G. Wahba, Y. Lin, and H. Zhang. Gacv for support vector machines. In A.J. Smola,
P.L. Bartlett, B. Sch?olkopf, and D. Schuurmans, editors, Advances in Large Margin
Classifiers, pages 297?311, Cambridge, MA, 2000. MIT Press.
[5] T. Hastie, R. Tibshirani, and J. Friedman. The Elements of Statistical Learning; Data
mining, Inference and Prediction. Springer Verlag, New York, 2001.
[6] Thorsten Joachims. Practical Advances in Kernel Methods ? Support Vector Learning, chapter Making large scale SVM learning practical. MIT Press, 1999. see
http://svmlight.joachims.org.
[7] Trevor Hastie, Saharon Rosset, Robert Tibshirani, and Ji Zhu. The entire regularization path for the support vector machine. Journal of Machine Learning Research,
(5):1391?1415, 2004.
[8] B. Efron, T. Hastie, I. Johnstone, and R.. Tibshirani. Least angle regression. Technical
report, Stanford University, 2002.
[9] Saharon Rosset and Ji Zhu.
Piecewise linear regularized solution
paths.
Technical
report,
Stanford
University,
2003.
http://www-stat.stanford.edu/?saharon/papers/piecewise.ps.
[10] Ji Zhu, Saharon Rosset, Trevor Hastie, and Robert Tibshirani. L1 norm support vector
machines. Technical report, Stanford University, 2003.
[11] Massimiliano Pontil and Alessandro Verri. Properties of support vector machines.
Neural Comput., 10(4):955?974, 1998.
[12] Shai Fine and Katya Scheinberg. Incas: An incremental active set method for svm.
Technical report, IBM Research Labs, Haifa, 2002.
[13] G. Cauwenberghs and T. Poggio. Incremental and decremental support vector machine learning. In Advances in Neural Information Processing Systems (NIPS*2000),
volume 13. MIT Press, Cambridge, MA, 2001.
[14] Christopher Diehl and Gert Cauwenberghs. Svm incremental learning, adaptation and
optimization. In Proceedings of the 2003 International Joint Conference on Neural
Networks, pages 2685?2690, 2003. Special series on Incremental Learning.
| 2713 |@word norm:4 mee:1 termination:1 simulation:2 decomposition:1 dramatic:1 carry:1 initial:4 configuration:1 series:1 hereafter:1 ours:1 com:1 must:4 happen:1 update:2 website:1 ith:1 org:1 zhang:1 height:1 along:4 surprised:1 fitting:3 inside:3 manner:1 inspired:1 xti:1 becomes:2 discover:1 notation:1 finding:1 guarantee:1 every:1 exactly:1 wrong:1 classifier:2 control:1 normally:1 unit:2 grant:2 subscript:1 path:22 interpolation:1 initialization:1 dut:1 katya:1 suggests:3 range:2 practical:2 yj:8 differs:1 procedure:2 pontil:1 radial:6 get:3 convenience:1 writing:2 restriction:1 equivalent:1 map:1 www:1 center:1 attention:1 convex:1 formulate:1 immediately:1 m2:1 continued:1 array:1 regarded:1 classic:1 gert:1 suppose:1 heavily:1 exact:3 programming:2 us:1 pa:1 element:3 updating:3 narrowing:1 role:1 tib:1 solved:1 decrease:6 movement:1 balanced:2 alessandro:1 environment:1 complexity:1 depend:1 solving:2 basis:1 gacv:1 easily:2 joint:1 represented:1 chapter:2 reparametrize:1 massimiliano:1 effective:4 outside:3 stanford:10 valued:1 solve:1 widely:1 larger:1 favor:1 statistic:3 itself:1 sequence:3 eigenvalue:5 maximal:1 adaptation:1 entered:1 achieve:1 intuitive:1 olkopf:2 empty:1 p:1 perfect:2 incremental:5 derive:1 illustrate:1 develop:3 stat:1 depending:1 b0:1 implemented:1 involves:1 implies:1 correct:1 attribute:1 shrunk:1 exploration:1 implementing:1 argued:1 sufficiently:1 around:1 exp:1 great:1 bj:2 achieves:2 early:1 largest:2 vice:1 establishes:1 rough:1 brought:1 mit:4 always:1 rather:1 avoid:2 focus:1 downdating:1 joachim:2 rank:5 check:1 pentium:1 attains:1 suppression:1 helpful:1 inference:1 entire:9 typically:2 initially:3 dual:1 orientation:2 colt:1 development:2 special:1 genuine:1 construct:3 saving:1 having:1 equal:1 unnecessarily:1 park:1 report:4 piecewise:16 employ:1 simultaneously:1 national:2 friedman:2 mining:1 mixture:5 primal:1 wiggly:1 experience:1 poggio:1 penalizes:1 haifa:1 srosset:1 minimal:2 column:2 soft:1 inca:1 cost:9 entry:1 predictor:1 examining:1 reported:1 connect:1 rosset:4 density:1 international:1 stay:2 linux:1 again:1 nm:1 derivative:1 return:1 sec:1 summarized:1 includes:1 coefficient:5 satisfy:1 break:1 closed:1 lab:1 cauwenberghs:2 start:2 shai:1 likewise:1 efficiently:1 sitting:1 correspond:1 generalize:1 emphasizes:2 trajectory:1 reach:1 trevor:4 definition:1 involved:1 dm:1 associated:1 mi:1 efron:1 blowing:1 response:1 verri:1 formulation:3 box:1 furthermore:1 just:1 implicit:2 smola:2 until:4 jerome:1 christopher:1 nonlinear:2 continuity:2 usa:2 effect:1 multiplier:3 true:1 regularization:8 hence:10 width:2 nuisance:1 noted:1 yorktown:1 criterion:5 ridge:1 saharon:5 l1:4 ji:4 exponentially:1 volume:1 occurred:2 versa:1 cambridge:2 stable:1 verlag:1 binary:1 watson:1 continue:1 yi:26 minimum:1 greater:1 employed:1 assisting:1 smoother:2 multiple:1 full:1 ii:1 technical:4 long:1 lin:1 prediction:1 involving:1 basic:1 regression:3 essentially:2 sometimes:1 kernel:19 achieved:1 fine:1 singular:1 sch:2 rest:1 eigenfunctions:2 subject:3 member:2 near:1 svmlight:2 easy:2 xj:8 fit:4 hastie:7 identified:1 lasso:2 wahba:1 reduce:1 simplifies:1 idea:1 whether:1 expression:1 bartlett:1 passed:1 penalty:1 york:1 constitute:1 ignored:1 amount:1 svms:5 http:2 affords:1 tibshirani:5 track:2 write:1 four:4 nevertheless:1 libsvm:3 run:1 inverse:3 angle:2 package:1 linger:1 guyon:1 separation:3 decision:1 prefer:1 completing:1 quadratic:8 occur:2 constraint:6 deficiency:1 alex:1 software:1 flat:1 min:7 separable:3 department:3 remain:1 smaller:1 ro1:1 lp:1 making:1 thorsten:1 equation:4 remains:1 scheinberg:1 slack:1 turn:1 needed:2 whichever:1 umich:1 available:1 eigen:1 rp:1 original:1 hinge:3 exploit:1 establish:4 implied:1 move:3 occurs:5 parametric:1 usual:1 diagonal:1 thank:1 argue:1 reason:2 toward:1 spanning:1 code:2 index:1 setup:2 robert:3 negative:1 unknown:1 vertical:1 observation:5 jizhu:1 truncated:1 situation:1 precise:1 annihilates:1 pair:2 required:1 extensive:1 optimized:1 narrow:1 nm2:1 established:1 boser:1 rougher:2 nip:1 beyond:2 critical:1 overlap:2 event:10 regularized:5 zhu:4 library:1 health:1 philadelphia:1 acknowledgement:1 loss:3 mixed:1 foundation:1 incurred:1 cn2:1 sufficient:1 editor:1 ibm:3 row:1 course:1 changed:1 penalized:2 supported:1 last:1 keeping:1 side:1 allow:2 institute:1 wide:3 johnstone:1 ghz:1 outermost:1 boundary:4 default:1 curve:2 dimension:1 gram:3 calculated:1 computes:1 author:3 collection:1 adaptive:1 far:2 decremental:1 gene:1 keep:1 kkt:1 active:1 xi:26 alternatively:1 table:3 terminate:1 nature:1 ca:2 diehl:1 schuurmans:1 complex:1 separator:2 did:1 main:1 linearly:1 rephrasing:1 join:1 position:1 wish:1 comput:1 lie:1 governed:1 young:1 down:2 r2:1 svm:16 cortes:1 burden:2 vapnik:2 effectively:1 margin:24 flavor:1 easier:1 michigan:1 lagrange:5 partially:1 springer:1 corresponds:1 ma:2 ann:1 change:7 hard:1 typical:1 infinite:1 called:1 total:2 pas:1 bernard:1 arbor:1 svd:1 m3:1 support:9 latter:1 arises:1 investigator:1 evaluate:1 |
1,887 | 2,714 | Harmonising Chorales by Probabilistic Inference
Moray Allan and Christopher K. I. Williams
School of Informatics, University of Edinburgh
Edinburgh EH1 2QL
[email protected], [email protected]
Abstract
We describe how we used a data set of chorale harmonisations composed
by Johann Sebastian Bach to train Hidden Markov Models. Using a probabilistic framework allows us to create a harmonisation system which
learns from examples, and which can compose new harmonisations. We
make a quantitative comparison of our system?s harmonisation performance against simpler models, and provide example harmonisations.
1
Introduction
Chorale harmonisation is a traditional part of the theoretical education of Western classical
musicians. Given a melody, the task is to create three further lines of music which will
sound pleasant when played simultaneously with the original melody. A good chorale
harmonisation will show an understanding of the basic ?rules? of harmonisation, which
codify the aesthetic preferences of the style. Here we approach chorale harmonisation as
a machine learning task, in a probabilistic framework. We use example harmonisations
to build a model of harmonic processes. This model can then be used to compose novel
harmonisations.
Section 2 below gives an overview of the musical background to chorale harmonisation.
Section 3 explains how we can create a harmonisation system using Hidden Markov Models. Section 4 examines the system?s performance quantitatively and provides example
harmonisations generated by the system. In section 5 we compare our system to related
work, and in section 6 we suggest some possible enhancements.
2
Musical Background
Since the sixteenth century, the music of the Lutheran church had been centred on the
?chorale?. Chorales were hymns, poetic words set to music: a famous early example is
Martin Luther?s ?Ein? feste Burg ist unser Gott?. At first chorales had only relatively simple melodic lines, but soon composers began to arrange more complex music to accompany
the original tunes. In the pieces by Bach which we use here, the chorale tune is taken generally unchanged in the highest voice, and three other musical parts are created alongside
it, supporting it and each other. By the eighteenth century, a complex system of rules had
developed, dictating what combinations of notes should be played at the same time or following previous notes. The added lines of music should not fit too easily with the melody,
but should not clash with it too much either. Dissonance can improve the music, if it is
resolved into a pleasant consonance.
Figure 1: Hidden state representations (a) for harmonisation, (b) for ornamentation.
The training and test chorales used here are divided into two sets: one for chorales in
?major? keys, and one for chorales in ?minor? keys. Major and minor keys are based around
different sets of notes, and musical lines in major and minor keys behave differently.
The representation we use to model harmonisations divides up chorales into discrete timesteps according to the regular beat underlying their musical rhythm. At each time-step we
represent the notes in the various musical parts by counting how far apart they are in terms
of all the possible ?semitone? notes.
3
3.1
Harmonisation Model
HMM for Harmonisation
We construct a Hidden Markov model in which the visible states are melody notes and the
hidden states are chords. A sequence of observed events makes up a melody line, and a
sequence of hidden events makes up a possible harmonisation for a melody line. We denote
the sequence of melody notes as Y and the harmonic motion as C, with yt representing the
melody at time t, and ct the harmonic state.
Hidden Markov Models are generative models: here we model how a visible melody line is
emitted by a hidden sequence of harmonies. This makes sense in musical terms, since we
can view a chorale as having an underlying harmonic structure, and the individual notes of
the melody line as chosen to be compatible with this harmonic state at each time step. We
will create separate models for chorales in major and minor keys, since these groups have
different harmonic structures.
For our model we divide each chorale into time steps of a single beat, making the assumption that the harmonic state does not change during a beat. (Typically there are three or
four beats in a bar.) We want to create a model which we can use to predict three further
notes at each of these time steps, one for each of the three additional musical lines in the
harmonisation.
There are many possible hidden state representations from which to choose. Here we represent a choice of notes by a list of pitch intervals. By using intervals in this way we
represent the relationship between the added notes and the melody at a given time step,
without reference to the absolute pitch of the melody note. These interval sets alone would
be harmonically ambiguous, so we disambiguate them using harmonic labels, which are
included in the training data set. Adding harmonic labels means that our hidden symbols
not only identify a particular chord, but also the harmonic function that the chord is serving. Figure 1(a) shows the representation used for some example notes. Here (an A major
chord) the alto, tenor and bass notes are respectively 4, 9, and 16 semitones below the
soprano melody. The harmonic label is ?T?, labelling this as functionally a ?tonic? chord.
Our representation of both melody and harmony distinguishes between a note which is
continued from the previous beat and a repeated note.
We make a first-order Markov assumption concerning the transition probabilities between
the hidden states, which represent choices of chord on an individual beat:
P (ct |ct?1 , ct?2 , . . . , c0 ) = P (ct |ct?1 ).
We make a similar assumption concerning emission probabilities to model how the observed event, a melody note, results from the hidden state, a chord:
P (yt |ct , . . . , c0 , yt?1 , . . . , y0 ) = P (yt |ct ).
In the Hidden Markov Models used here, the ?hidden? states of chords and harmonic symbols are in fact visible in the data during training. This means that we can learn transition
and emission probabilities directly from observations in our training data set of harmonisations. We use additive smoothing (adding 0.01 to each bin) to deal with zero counts in the
training data.
Using a Hidden Markov Model framework allows us to conduct efficient inference over
our harmonisation choices. In this way our harmonisation system will ?plan? over an entire
harmonisation rather than simply making immediate choices based on the local context.
This means, for example, that we can hope to compose appropriate ?cadences? to bring our
harmonisations to pleasant closes rather than finishing abruptly.
Given a new melody line, we can use the Viterbi algorithm to find the most likely state
sequence, and thus harmonisation, given our model. We can also provide alternative harmonisations by sampling from the posterior [see 1, p. 156], as explained below.
3.2
Sampling Alternative Harmonisations
Using ?t?1 (j), the probability of seeing the observed events of a sequence up to time t ? 1
and finishing in state j, we can calculate the probability of seeing the first t ? 1 events,
finishing in state j, and then transitioning to state k at the next step:
P (y0 , y1 , . . . , yt?1 , ct?1 = j, ct = k) = ?t?1 (j)P (ct = k|ct?1 = j).
We can use this to calculate ?t (j|k), the probability that we are in state j at time t ? 1 given
the observed events up to time t ? 1, and given that we will be in state k at time t:
?t?1 (j)P (ct = k|ct?1 = j)
?t (j|k) = P (ct?1 = j|y0 , y1 , . . . , yt?1 , ct = k) = P
.
l ?t?1 (l)P (ct = k|ct?1 = l)
To sample from P (C|Y ) we first choose the final state by sampling from its probability
distribution according to the model:
?T (j)
P (cT = j|y0 , y1 , . . . , yT ) = P
.
l ?T (l)
Once we have chosen a value for the final state cT , we can use the variables ?t (j|k) to
sample backwards through the sequence:
P (ct = j|y0 , y1 , . . . , yT , ct+1 ) = ?t+1 (j|ct+1 ).
3.3
HMM for Ornamentation
The chorale harmonisations produced by the Hidden Markov Model described above harmonise the original melody according to beat-long time steps. Chorale harmonisations are
Table 1: Comparison of predictive power achieved by different models of harmonic sequences on training and test data sets (nats).
? T1
? T1
? T1
? T1
ln
PP (C|Y )
P ln P (ct |yt )
P ln P (ct |ct?1 )
ln P (ct )
Training (maj)
Test (maj)
Training (min)
Training (min)
2.56
3.00
5.41
6.43
4.90
3.22
7.08
7.61
2.66
3.52
5.50
6.57
5.02
4.33
7.21
7.84
not limited to this rhythmic form, so here we add a secondary ornamentation stage which
can add passing notes to decorate these harmonisations. Generating a harmonisation and
adding the ornamentation as a second stage greatly reduces the number of hidden states
in the initial harmonisation model: if we went straight to fully-ornamented hidden states
then the data available to us concerning each state would be extremely limited. Moreover,
since the passing notes do not change the harmonic structure of a piece but only ornament
it, adding these passing notes after first determining the harmonic structure for a chorale is
a plausible compositional process.
We conduct ornamentation by means of a second Hidden Markov Model. The notes added
in this ornamentation stage generally smooth out the movement between notes in a line of
music, so we set up the visible states in terms of how much the three harmonising musical
lines rise or fall from one time-step to the next. The hidden states describe ornamentation
of this motion in terms of the movement made by each part during the time step, relative
to its starting pitch. This relative motion is described at a time resolution four times as fine
as the harmonic movement. On the first of the four quarter-beats we always leave notes
as they were, so we have to make predictions only for the final three quarter-beats. Figure
1(b) shows an example of the representation used. In this example, the alto and tenor lines
remain at the same pitch for the second quarter-beat as they were for the first, and rise by
two semitones for the third and fourth quarter-beats, so are both represented as ?0,0,2,2?,
while the bass line does not change pitch at all, so is represented as ?0,0,0,0?.
4
Results
Our training and test data are derived from chorale harmonisations by Johann Sebastian
Bach.1 These provide a relatively large set of harmonisations by a single composer, and are
long established as a standard reference among music theorists. There are 202 chorales in
major keys of which 121 were used for training and 81 used for testing; and 180 chorales
in minor keys (split 108/72).
Using a probabilistic framework allows us to give quantitative answers to questions about
the performance of the harmonisation system. There are many quantities we could compute, but here we will look at how high a probability the model assigns to Bach?s own
harmonisations given the respective melody lines. We calculate average negative log probabilities per symbol, which describe how predictable the symbols are under the model.
These quantities provide sample estimates of cross-entropy. Whereas verbal descriptions
of harmonisation performance are unavoidably vague and hard to compare, these figures
allow our model?s performance to be directly compared with that of any future probabilistic
harmonisation system.
Table 1 shows the average negative log probability per symbol of Bach?s chord symbol
1
We used a computer-readable edition of Bach?s chorales downloaded from ftp://i11ftp.
ira.uka.de/pub/neuro/dominik/midifiles/bach.zip
Figure 2: Most likely harmonisation under our model of chorale K4, BWV 48
Figure 3: Most likely harmonisation under our model of chorale K389, BWV 438
sequences given their respective melodic symbol sequences, ? T1 ln P (C|Y ), on training
and test data sets of chorales in major and minor keys. As a comparison we give analogous negative log probabilities
for a model predicting chord states from their respective
P
ln P (ct |yt ), for a simple Markov chain between the chord states,
melody notes, ? T1
P
? T1
ln P (ct |ct?1 ), and for a model which assumes that the chord states are indepenP
dently drawn, ? T1
ln P (ct ). The Hidden Markov Model here has 5046 hidden chord
states and 58 visible melody states.
The Hidden Markov Model finds a better fit to the training data than the simpler models:
to choose a good chord for a particular beat we need to take into account both the melody
note on that beat and the surrounding chords. Even the simplest model of the data, which
assumes that each chord is drawn independently, performs worse on the test data than the
training data, showing that we are suffering from sparse data. There are many chords, chord
to melody note emissions, and especially chord to chord transitions, that are seen in the test
data but never occur in the training data. The models? performance with unseen data could
be improved by using a more sophisticated smoothing method, for example taking into
account the overall relative frequencies of harmonic symbols when assigning probabilities
to unseen chord transitions. However, this lower performance with unseen test data is not
a problem for the task we approach here, of generating new harmonisations, as long as we
can learn a large enough vocabulary of events from the training data to be able to find good
harmonisations for new chorale melodies.
Figures 2 and 3 show the most likely harmonisations under our model for two short
chorales. The system has generated reasonable harmonisations. We can see, for example,
passages of parallel and contrary motion between the different parts. There is an appropriate harmonic movement through the harmonisations, and they come to plausible cadences.
The generated harmonisations suffer somewhat from not taking into account the flow of
the individual musical lines which we add. There are large jumps, especially in the bass
line, more often than is desirable ? the bass line suffers most since has the greatest variance
with respect to the soprano melody. This excessive jumping also feeds through to reduce
the performance of the ornamentation stage, creating visible states which are unseen in the
training data. The model structure means that the most likely harmonisation leaves these
states unornamented. Nevertheless, where ornamentation has been added it fits with its
context and enhances the harmonisations.
The authors will publish further example harmonisations, including MIDI files, online at
http://www.tardis.ed.ac.uk/?moray/harmony/.
5
Relationship to previous work
Even while Bach was still composing chorales, music theorists were catching up with musical practice by writing treatises to explain and to teach harmonisation. Two famous examples, Rameau?s Treatise on Harmony [2] and the Gradus ad Parnassum by Fux [3],
show how musical style was systematised and formalised into sets of rules. The traditional
formulation of harmonisation technique in terms of rules suggests that we might create an
automatic harmonisation system by finding as many rules as we can and encoding them as
a consistent set of constraints. Pachet and Roy [4] provide a good overview of constraintbased harmonisation systems. For example, one early system [5] takes rules from Fux and
assigns penalties according to the seriousness of each rule being broken. This system then
conducts a modified best-first search to produce harmonisations. Using standard constraintsatisfaction techniques for harmonisation is problematic, since the space and time needs of
the solver tend to rise extremely quickly with the length of the piece.
Several systems have applied genetic programming techniques to harmonisation, for example McIntyre [6]. These are similar to the constraint-based systems described above,
but instead of using hard constraints they encode their rules as a fitness function, and try
to optimise that function by evolutionary techniques. Phon-Amnuaisuk and Wiggins [7]
are reserved in their assessment of genetic programming for harmonisation. They make a
direct comparison with an ordinary constraint-based system, and conclude that the performance of each system is related to the amount of knowledge encoded in it rather than the
particular technique it uses. In their comparison the ordinary constraint-based system actually performs much better, and they argue that this is because it possesses implicit control
knowledge which the system based on the genetic algorithm lacks.
Even if they can be made more efficient, these rule-based systems do not perform the full
task of our harmonisation system. They take a large set of rules written by a human and
attempt to find a valid solution, whereas our system learns its rules from examples.
Hild et al. [8] use neural networks to harmonise chorales. Like the Hidden Markov Models
in our system, these neural networks are trained using example harmonisations. However,
while two of their three subtasks use only neural networks trained on example harmonisations, their second subtask, where chords are chosen to instantiate more general harmonies,
includes constraint satisfaction. Rules written by a human penalise undesirable combinations of notes, so that they will be filtered out when the best chord is chosen from all
those compatible with the harmony already decided. In contrast, our model learns all its
harmonic ?rules? from its training data.
Ponsford et al. [9] use n-gram Markov models to generate harmonic structures. Unlike in
chorale harmonisation, there is no predetermined tune with which the harmonies need to fit.
The data set they use is a selection of 84 saraband dances, by 15 different seventeen-century
French composers. An automatically annotated corpus is used to train Markov models
using contexts of different lengths, and the weighted sum of the probabilities assigned by
these models used to predict harmonic movement. Ponsford et al. create new pieces first by
random generation from their models, and secondly by selecting those randomly-generated
pieces which match a given template. Using templates gives better results, but the great
majority of randomly-generated pieces will not match the template and so will have to be
discarded. Using a Hidden Markov Model rather than simple n-grams allows this kind of
template to be included in the model as the visible state of the system: the chorale tunes
in our system can be thought of as complex templates for harmonisations. Ponsford et
al. note that even with their longest context length, the cadences are poor. In our system
the ?planning? ability of Hidden Markov Models, using the combination of chords and
harmonic labels encoded in the hidden states, produces cadences which bring the chorale
tunes to harmonic closure.
This paper stems from work described in the first author?s MSc thesis [10] carried out in
2002. We have recently become aware that similar work has been carried out independently
in Japan by a team led by Prof S. Sagayama [11, 12]. To our knowledge this work has
been published only in Japanese2 . The basic frameworks are similar, but there are several
differences. First, their system only describes the harmonisation in terms of the harmonic
label (e.g. T for tonic) and does not fully specify the voicing of the three harmony lines or
ornamentation. Secondly, they do not give a quantitative evaluation of the harmonisations
produced as in our Table 1. Thirdly, in [12] a Markov model on blocks of chord sequences
rather than on individual chords is explored.
6
Discussion
Using the framework of probabilistic influence allows us to perform efficient inference to
generate new chorale harmonisations, avoiding the computational scaling problems suffered by constraint-based harmonisation systems. We described above neural network
and genetic algorithm techniques which were less compute-intensive than straightforward
constraint satisfaction, but the harmonisation systems using these techniques retain a preprogrammed knowledge base, whereas our model is able to learn its harmonisation constraints from training data.
Different forms of graphical model would allow us to take into account more of the dependencies in harmonisation. For example, we could use a higher-order Markov structure,
although this by itself would be likely to greatly increase the problems already seen here
with sparse data. An alternative might be to use an Autoregressive Hidden Markov Model
[13], which models the transitions between visible states as well as the hidden state transitions modelled by an ordinary Hidden Markov Model.
Not all of Bach?s chorale harmonisations are in the same style. Some of his harmonisations
are intentionally complex, and others intentionally simple. We could improve our harmonisations by modelling this stylistic variation, either manually annotating training chorales
according to their style or by training a mixture of HMMs.
As we only wish to model the hidden harmonic state given the melody, rather than construct
a full generative model of the data, Conditional Random Fields (CRFs) [14] provide a
related but alternative framework. However, note that training such models (e.g. using
iterative scaling methods) is more difficult than the simple counting methods that can be
applied to the HMM case. On the other hand the use of the CRF framework would have
2
We thank Yoshinori Shiga for explaining this work to us.
some advantages, in that additional features could be incorporated. For example, we might
be able to make better predictions by taking into account the current time step?s position
within its musical bar. Music theory recognises a hierarchy of stressed beats within a bar,
and harmonic movement should correlated with these stresses. The ornamentation process
especially might benefit from a feature-based approach.
Our system described above only considers chords as sets of intervals, and thus does not
have a notion of the key of a piece (other than major or minor). However, voices have a
preferred range and thus the notes that should be used do depend on the key, so the key
signature could also be used as a feature in a CRF. Taking into account the natural range of
each voice would prevent the bass line from descending too low and keep the three parts
closer together. In general more interesting harmonies result when musical lines are closer
together and their movements are more constrained. Another dimension that could be
explored with CRFs would be to take into account the words of the chorales, since Bach?s
own harmonisations are affected by the properties of the texts as well as of the melodies.
Acknowledgments
MA gratefully acknowledges support through a research studentship from Microsoft Research Ltd.
References
[1] R. Durbin, S. R. Eddy, A. Krogh, and G. Mitchison. Biological sequence analysis. Cambridge
University Press, 1998.
[2] J.-P. Rameau. Trait?e de l?Harmonie reduite a` ses principes naturels. Paris, 1722.
[3] J. J. Fux. Gradus ad Parnassum. Vienna, 1725.
[4] F. Pachet and P. Roy. Musical harmonization with constraints: A survey. Constraints, 6(1):
7?19, 2001.
[5] B. Schottstaedt.
CCRMA, 1989.
Automatic species counterpoint.
Technical report, Stanford University
[6] R. A. McIntyre. Bach in a box: The evolution of four-part baroque harmony using the genetic
algorithm. In Proceedings of the IEEE Conference on Evolutionary Computation, 1994.
[7] S. Phon-Amnuaisuk and G. A. Wiggins. The four-part harmonisation problem: a comparison
between genetic algorithms and a rule-based system. In Proceedings of the AISB?99 Symposium
on Musical Creativity, 1999.
[8] H. Hild, J. Feulner, and W. Menzel. HARMONET: A neural net for harmonizing chorales in
the style of J.S. Bach. In R.P. Lippman, J.E. Moody, and D.S. Touretzky, editors, Advances in
Neural Information Processing 4, pages 267?274. Morgan Kaufmann, 1992.
[9] D. Ponsford, G. Wiggins, and C. Mellish. Statistical learning of harmonic movement. Journal
of New Music Research, 1999.
[10] M. M. Allan. Harmonising Chorales in the Style of Johann Sebastian Bach. Master?s thesis,
School of Informatics, University of Edinburgh, 2002.
[11] T. Kawakami. Hidden Markov Model for Automatic Harmonization of Given Melodies. Master?s thesis, School of Information Science, JAIST, 2000. In Japanese.
[12] K. Sugawara, T. Nishimoto, and S. Sagayama. Automatic harmonization for melodies based on
HMMs including note-chain probability. Technical Report 2003-MUS-53, Acoustic Society of
Japan, December 2003. In Japanese.
[13] P. C. Woodland. Hidden Markov Models using vector linear prediction and discriminative
output distributions. In Proc ICASSP, volume I, pages 509?512, 1992.
[14] J. D. Lafferty, A. McCallum, and F. C. N. Pereira. Conditional Random Fields: probabilistic
models for segmenting and labeling sequence data. In Proc ICML, pages 282?289, 2001.
| 2714 |@word c0:2 closure:1 initial:1 pub:1 selecting:1 genetic:6 feulner:1 clash:1 current:1 assigning:1 written:2 additive:1 visible:8 predetermined:1 alone:1 generative:2 leaf:1 instantiate:1 mccallum:1 short:1 filtered:1 provides:1 preference:1 simpler:2 harmonically:1 direct:1 become:1 symposium:1 compose:3 allan:3 planning:1 automatically:1 solver:1 underlying:2 moreover:1 alto:2 what:1 kind:1 developed:1 finding:1 quantitative:3 penalise:1 uk:3 control:1 segmenting:1 t1:8 local:1 encoding:1 might:4 suggests:1 hmms:2 limited:2 range:2 sugawara:1 decided:1 acknowledgment:1 testing:1 practice:1 block:1 aisb:1 lippman:1 thought:1 word:2 regular:1 seeing:2 melodic:2 suggest:1 close:1 undesirable:1 selection:1 context:4 pachet:2 writing:1 influence:1 descending:1 www:1 nishimoto:1 eighteenth:1 musician:1 yt:10 williams:2 straightforward:1 starting:1 independently:2 crfs:2 survey:1 resolution:1 assigns:2 rule:14 examines:1 continued:1 his:1 century:3 notion:1 variation:1 analogous:1 hierarchy:1 programming:2 us:1 roy:2 observed:4 calculate:3 bass:5 went:1 movement:8 highest:1 chord:27 subtask:1 predictable:1 broken:1 mu:1 nats:1 preprogrammed:1 signature:1 trained:2 depend:1 predictive:1 vague:1 easily:1 resolved:1 icassp:1 differently:1 various:1 soprano:2 represented:2 surrounding:1 train:2 describe:3 labeling:1 encoded:2 stanford:1 plausible:2 s:1 annotating:1 ability:1 unseen:4 itself:1 final:3 online:1 sequence:13 advantage:1 net:1 formalised:1 unavoidably:1 sixteenth:1 description:1 enhancement:1 produce:2 generating:2 leave:1 ftp:1 ac:3 school:3 minor:7 krogh:1 come:1 annotated:1 human:2 melody:28 education:1 explains:1 bin:1 creativity:1 baroque:1 harmonization:3 biological:1 secondly:2 hild:2 around:1 great:1 viterbi:1 predict:2 major:8 arrange:1 early:2 proc:2 harmony:10 label:5 create:7 weighted:1 hope:1 always:1 modified:1 rather:6 harmonizing:1 gradus:2 ira:1 encode:1 derived:1 emission:3 finishing:3 longest:1 modelling:1 greatly:2 contrast:1 sense:1 inference:3 typically:1 entire:1 shiga:1 hidden:33 overall:1 among:1 plan:1 smoothing:2 constrained:1 field:2 construct:2 once:1 having:1 dissonance:1 sampling:3 never:1 aware:1 manually:1 harmonet:1 look:1 icml:1 excessive:1 future:1 others:1 cadence:4 quantitatively:1 report:2 distinguishes:1 seriousness:1 randomly:2 composed:1 simultaneously:1 mellish:1 individual:4 fitness:1 microsoft:1 attempt:1 evaluation:1 mixture:1 ornamentation:11 chain:2 closer:2 respective:3 jumping:1 conduct:3 divide:2 catching:1 theoretical:1 ordinary:3 parnassum:2 menzel:1 too:3 dependency:1 answer:1 phon:2 retain:1 probabilistic:7 informatics:2 together:2 quickly:1 moody:1 thesis:3 choose:3 worse:1 creating:1 style:6 japan:2 account:7 de:2 centred:1 includes:1 ad:2 piece:7 view:1 try:1 parallel:1 musical:16 variance:1 reserved:1 kaufmann:1 identify:1 modelled:1 famous:2 produced:2 straight:1 published:1 explain:1 suffers:1 touretzky:1 sebastian:3 ed:3 decorate:1 against:1 pp:1 frequency:1 intentionally:2 seventeen:1 knowledge:4 eddy:1 sophisticated:1 actually:1 feed:1 higher:1 specify:1 improved:1 formulation:1 box:1 stage:4 implicit:1 msc:1 hand:1 christopher:1 assessment:1 western:1 lack:1 french:1 kawakami:1 evolution:1 assigned:1 codify:1 deal:1 semitone:3 during:3 ambiguous:1 rhythm:1 stress:1 crf:2 performs:2 motion:4 bring:2 passage:1 harmonic:27 novel:1 mcintyre:2 recently:1 began:1 quarter:4 overview:2 volume:1 thirdly:1 functionally:1 trait:1 theorist:2 cambridge:1 automatic:4 gratefully:1 had:3 treatise:2 add:3 base:1 posterior:1 own:2 apart:1 seen:2 morgan:1 additional:2 somewhat:1 zip:1 full:2 sound:1 desirable:1 reduces:1 stem:1 smooth:1 technical:2 match:2 bach:13 long:3 cross:1 divided:1 concerning:3 pitch:5 prediction:3 basic:2 neuro:1 publish:1 represent:4 achieved:1 background:2 want:1 fine:1 whereas:3 interval:4 uka:1 suffered:1 unlike:1 posse:1 file:1 accompany:1 tend:1 december:1 contrary:1 flow:1 lafferty:1 emitted:1 counting:2 backwards:1 aesthetic:1 split:1 enough:1 fit:4 timesteps:1 reduce:1 intensive:1 ltd:1 penalty:1 abruptly:1 suffer:1 passing:3 compositional:1 dictating:1 generally:2 woodland:1 pleasant:3 tune:5 amount:1 simplest:1 http:1 generate:2 problematic:1 per:2 serving:1 dently:1 discrete:1 affected:1 ist:1 key:11 group:1 four:5 nevertheless:1 drawn:2 k4:1 prevent:1 sum:1 fourth:1 master:2 reasonable:1 stylistic:1 scaling:2 ct:31 played:2 durbin:1 occur:1 constraint:11 min:2 extremely:2 martin:1 relatively:2 according:5 combination:3 poor:1 remain:1 describes:1 y0:5 making:2 explained:1 counterpoint:1 taken:1 ln:8 count:1 available:1 appropriate:2 voicing:1 alternative:4 voice:3 original:3 assumes:2 burg:1 graphical:1 readable:1 vienna:1 music:11 build:1 especially:3 prof:1 classical:1 society:1 unchanged:1 eh1:1 added:4 question:1 quantity:2 already:2 recognises:1 traditional:2 enhances:1 evolutionary:2 separate:1 thank:1 hmm:3 majority:1 argue:1 considers:1 length:3 relationship:2 ql:1 difficult:1 teach:1 negative:3 rise:3 perform:2 ornament:1 observation:1 markov:23 discarded:1 behave:1 supporting:1 beat:14 immediate:1 fux:3 tonic:2 team:1 incorporated:1 y1:4 wiggins:3 subtasks:1 paris:1 acoustic:1 established:1 able:3 bar:3 alongside:1 below:3 including:2 optimise:1 power:1 event:7 greatest:1 satisfaction:2 natural:1 predicting:1 representing:1 chorale:40 improve:2 jaist:1 created:1 church:1 maj:2 carried:2 acknowledges:1 text:1 understanding:1 determining:1 relative:3 fully:2 generation:1 interesting:1 downloaded:1 consistent:1 editor:1 compatible:2 soon:1 verbal:1 allow:2 fall:1 template:5 taking:4 explaining:1 absolute:1 rhythmic:1 sparse:2 edinburgh:3 benefit:1 studentship:1 dimension:1 vocabulary:1 transition:6 valid:1 gram:2 autoregressive:1 author:2 made:2 jump:1 far:1 midi:1 preferred:1 keep:1 corpus:1 conclude:1 discriminative:1 mitchison:1 tenor:2 search:1 iterative:1 table:3 disambiguate:1 learn:3 composing:1 composer:3 complex:4 japanese:2 sagayama:2 edition:1 repeated:1 suffering:1 ein:1 position:1 pereira:1 wish:1 dominik:1 third:1 learns:3 transitioning:1 showing:1 symbol:8 unser:1 list:1 explored:2 adding:4 labelling:1 entropy:1 led:1 simply:1 likely:6 poetic:1 ma:1 conditional:2 change:3 hard:2 included:2 specie:1 secondary:1 principe:1 support:1 stressed:1 johann:3 dance:1 avoiding:1 correlated:1 |
1,888 | 2,715 | Theories Of Access Consciousness
Michael D. Colagrosso
Department of Computer Science
Colorado School of Mines
Golden, CO 80401 USA
[email protected]
Michael C. Mozer
Institute of Cognitive Science
University of Colorado
Boulder, CO 80309 USA
[email protected]
Abstract
Theories of access consciousness address how it is that some mental states but
not others are available for evaluation, choice behavior, and verbal report. Farah,
O?Reilly, and Vecera (1994) argue that quality of representation is critical; Dehaene, Sergent, and Changeux (2003) argue that the ability to communicate representations is critical. We present a probabilistic information transmission or
PIT model that suggests both of these conditions are essential for access consciousness. Having successfully modeled data from the repetition priming literature in the past, we use the PIT model to account for data from two experiments
on subliminal priming, showing that the model produces priming even in the absence of accessibility and reportability of internal states. The model provides a
mechanistic basis for understanding the dissociation of priming and awareness.
Philosophy has made many attempts to identify distinct aspects of consciousness. Perhaps
the most famous effort is Block?s (1995) delineation of phenomenal and access consciousness. Phenomenal consciousness has to do with ?what it is like? to experience chocolate
or a pin prick. Access consciousness refers to internal states whose content is ?(1) inferentially promiscuous, i.e., poised to be used as a premise in reasoning, (2) poised for control
of action, and (3) poised for rational control of speech.? (p. 230) The scientific study of consciousness has exploded in the past six years, and an important catalyst for this explosion
has been the decision to focus on the problem of access consciousness: how is it that some
mental states but not others become available for evaluation, choice behavior, verbal report,
and storage in working memory. Another reason for the recent explosion of consciousness
research is the availability of functional imaging techniques to explore differences in brain
activation between conscious and unconscious states, as well as the development of clever
psychological experiments that show that a stimulus that is not consciously perceived can
nonetheless influence cognition, which we describe shortly.
1
Subliminal Priming
The phenomena we address utilize an experimental paradigm known as repetition priming.
Priming refers to an improvement in efficiency in processing a stimulus item as a result
of previous exposure to the item. Efficiency is defined in terms of shorter response times,
lower error rates, or both. A typical long-term perceptual priming experiment consists of
a study phase during which participants are asked to read aloud a list of words, and a test
phase during which participants must name or categorize a series of words, presented one
at a time. Reaction time is lower and/or accuracy is higher for test words that were also on
the study list. Repetition priming occurs without strategic effort on the part of participants,
and therefore appears to be a low level mechanism of learning, which likely serves as the
mechanism underlying the refinement of cognitive skills with practice.
In traditional studies, priming is supraliminal?the prime is consciously perceived. In the
studies we model here, primes are subliminal. Subliminal priming addresses fundamental
issues concerning conscious access: How is it that a word or image that cannot be identified,
detected, or even discriminated in forced choice can nonetheless influence the processing
of a subsequent stimulus word? Answering this question in a computational framework
would be a significant advance toward understanding the nature of access consciousness.
2
Models of Conscious and Unconscious Processing
In contrast to the wealth of experimental data, and the large number of speculative and
philosophical papers on consciousness, concrete computational models are rare. The domain of consciousness is particularly ripe for theoretical perspectives, because it is a significant contribution to simply provide an existence proof of a mechanism that can explain
specific experimental data. Ordinarily, a theorist faces skepticism when presenting a model;
it often seems that hundreds of alternative, equally plausible accounts must exist. However,
when addressing data deemed central to issues of consciousness, simply providing a concrete handle on the phenomena serves to demystify consciousness and bring it into the
realm of scientific understanding.
We are familiar with only three computational models that address specific experimental
data in the domain of consciousness. We summarize these models, and then present a
novel model and describe its relationship to the previous efforts. Farah, O?Reilly, and Vecera (1994) were the first to model specific phenomena pertaining to consciousness in a
computational framework. The phenomena involve prosopagnosia, a deficit of overt face
recognition following brain damage. Nonetheless, prosopagnosia patients exhibit residual
covert recognition by a variety of tests. For example, when patients are asked to categorize
names as famous or nonfamous, their response times are faster to a famous name when the
name is primed by a picture of a semantically related face (e.g., the name ?Bill Clinton?
when preceded by a photograph of Hillary), despite the fact that they could not identify the
related face. Farah et al. model face recognition in a neural network, and show that when
the network is damaged, it loses the ability to perform tasks requiring high fidelity representations (e.g., identification) but not tasks requiring only coarse information (e.g., semantic
priming). They argue that conscious perception is associated with a certain minimal quality
of representation.
Dehaene and Naccache (2001) outline a framework based on Baars? (1989) notion of conscious states as residing in a global workspace. They describe the workspace as a ?distributed neural system...with long-distance connectivity that can potentially interconnect
multiple specialized brain areas in a coordinated, though variable manner.? (p. 13) Dehaene, Sergent, and Changeaux (2003) implement this framework in a complicated architecture of integrate-and-fire neurons and show that the model can qualitatively account for
the attentional blink phenomenon. The attentional blink is observed in experiments where
participants are shown a rapid series of stimuli, which includes two targets (T1 and T2).
If T2 appears shortly after T1, the ability to report T2 drops, as if attention is distracted.
Dehane et al. explain this phenomenon as follows. When T1 is presented, its activation
propagates to frontal cortical areas (the global workspace). Feedback connections lead to
a resonance between frontal and posterior areas, which strengthen T1 but block T2 from
entering the workspace. If the T1-T2 lag is sufficiently great, habituation of T1 sufficiently
weakens the representation such that T2 can enter the workspace and suppress T1. In this
account, conscious access is achieved via resonance between posterior and frontal areas.
Although the Farah et al. and Dehaene et al. models might not seem to have much in
common, they both make claims concerning what is required to achieve functional connectivity between perceptual and response systems. Farah et al. focus on aspects of the
representation; Dehaene et al. focus on a pathway through which representations can be
communicated. These two aspects are not incompatible, and in fact, a third model incorporates both. Mathis and Mozer (1996) describe an architecture with processing modules for
perceptual and response processes, implemented as attractor neural nets. They argue that in
order for a representation in some perceptual module to be assured of influencing a response
module, (a) it must have certain characteristics?temporal persistence and well-formedness?
which is quite similar to Farah et al.?s notion of quality, and (b) the two modules must be
interconnected?which is the purpose of Dehaene et al.?s global workspace. The model has
two limitations that restrict its value as a contemporary account of conscious access. First,
it addressed classical subliminal priming data, but more reliable data has recently been
reported. Second, like the other two models, Mathis and Mozer used a complex neural
network architecture with arbitrary assumptions built in, and the sensitivity of the model?s
behavior to these assumptions is far from clear. In this paper, we present a model that embodies the same assumptions as Mathis and Mozer, but overcomes its two limitations, and
explains subliminal-priming data that has yet to be interpreted via a computational model.
3
The Probabilistic Information Transmission (PIT) Framework
Our model is based on the probabilistic information transmission or PIT framework of
Mozer, Colagrosso, and Huber (2002, 2003). The framework characterizes the transmission
of information from perceptual to response systems, and how the time course of information transmission changes with experience (i.e., priming). Mozer et al. used this framework
to account for a variety of facilitation effects from supraliminal repetition priming.
The framework describes cognition in terms of a collection of information-processing pathways, and supposes that any act of cognition involves coordination among multiple pathways. For example, to model a letter-naming task where a letter printed in upper or lower
case is presented visually and the letter must be named, the framework would assume a
perceptual pathway that maps the visual input to an identity representation, and a response
pathway that maps a identity representation to a naming response. The framework is formalized as a probabilistic model: the pathway input and output are random variables and
microinference in a pathway is carried out by Bayesian belief revision.
The framework captures the time course of information processing for a single experimental trial. To elaborate, consider a pathway whose input at time t is a discrete random
variable, denoted X(t), which can assume values x1 , x2 , x3 , . . . , xnx corresponding to alternative input states. Similarly, the output of the pathway at time t is a discrete random
variable, denoted Y (t), which can assume values y1 , y2 , y3 , . . . , yny . For example, in the
letter-naming task, the input to the perceptual pathway would be one of nx = 52 visual
patterns corresponding to the upper- and lower-case letters of the alphabet, and the output
is one of ny = 26 letter identities. To present a particular input alternative, say xi , to the
model for T time steps, we specify X(t) = xi for t = 1 . . . T , and allow the model to
compute P(Y (t) | X(1) . . . X(t)).
A pathway is modeled as a dynamic Bayes network; the minimal version of the model
used in the present simulations is simply a hidden Markov model, where the X(t) are
observations and the Y (t) are inferred state (see Figure 1a). In typical usage, an HMM
is presented with a sequence of distinct inputs, whereas we maintain the same input for
many successive time steps; and an HMM transitions through a sequence of distinct hidden
states, whereas we attempt to converge with increasing confidence on a single state.
Figure 1b illustrates the time course of inference in a single pathway with 52 input and
26 output alternatives and two-to-one associations. The solid line in the Figure shows, as
a function of time t, P(Y (t) = yi | X(1) = x2i . . . X(t) = x2i ), i.e., the probability
that input i (say, the visual pattern of an upper case O) will produce its target output (the
letter identity). Evidence for the target output accumulates gradually over time, yielding a
speed-accuracy curve that relates the number of iterations to the accuracy of identification.
Y0
Y1
X1
Y2
X2
P(Output)
1
Y3
X3
(a)
(b)
0.8
0.6
O
0.4
0.2
0
Q
Time
Figure 1: (a) basic pathway architecture?a hidden Markov model; (b) time course of inference in a
pathway when the letter O is presented, causing activation of both O and the visually similar Q.
The exact shape of the speed-accuracy curve?the pathway dynamics?are determined by
three probability distributions, which embody the knowledge and past experience of the
model. First, P(Y (0)) is the prior distribution over outputs in the absence of any information about the input. Second, P(Y (t) | Y (t ? 1)) characterizes how the pathway output
evolves over time. We assume the transition probability matrix serves as a memory with
diffusion, i.e., P(Y (t) = yi |Y (t ? 1) = yj ) = (1 ? ?)?ij + ?P(Y (0) = yi ), where ? is the
diffusion constant and ?ij is the Kronecker delta. Third, P(X(t) | Y (t)) characterizes the
strength of association between inputs and outputs. The greater the association strength,
the more rapidly that information about X will be communicated
to Y . We parameterize
P
this distribution as P(X(t) = xi |Y (t) = yj ) ? 1 + k ?ik ?kj , where ?ij indicates the
frequency of experience with the association between states xi and yj , and ?ik specifies the
similarity between states xi and xk . (Although the representation of states is localist, the ?
terms allow us to design in the similarity structure inherent in a distributed representation.)
These association strengths are highly constrained by the task structure and the similarity
structure and familiarity of the inputs.
Fundamental to the framework is the assumption that with each experience, a pathway becomes more efficient at processing an input. Efficiency is reflected by a shift in the speedaccuracy curve to the left. In Mozer, Colagrosso, and Huber (2002, 2003), we propose two
distinct mechanisms to model phenomena of supraliminal priming. First, the association
frequencies, ?ij , are increased following a trial in which xi leads to activation of yj , resulting in more efficient transmission of information, corresponding to an increased slope
of the solid line in Figure 1b. The increase is Hebbian, based on the maximum activation
achieved by xi and yj : ??ij = ? maxt P(X(t) = xi )P(Y (t) = yj ), where ? is a step size.
Second, the priors, which serve as a model of the environment, are increased to indicate a
greater likelihood of the same output occurring again in the future. In modeling data from
supraliminal priming, we found that the increases to association frequencies are long lasting, but the increases to the priors decay over the course of a few minutes or a few trials.
As a result, the prior updating does not play into the simulation we report here; we refer
the reader to Mozer, Colagrosso, and Huber (2003) for details.
4
Access Consciousness and PIT
We have described the operation of a single pathway, but to model any cognitive task,
we require a series of pathways in cascade. For a simple choice task, we use a percpetual pathway cascaded to a response pathway. The interconnection between the pathways
is achieved by copying the output of the perceptual pathway, Y p (t), to the input of the
response pathway, X r (t), at each time t.
This multiple-pathway architecture allows us to characterize the notion of access consciousness. Considering the output of the perceptual pathway, access is achieved when:
(1) the output representation is sufficient to trigger the correct behavior in the response
pathway, and (2) the perceptual and response pathways are functionally interconnected. In
more general terms, access for a perceptual pathway output requires that these two condi-
tions be met not just for a specific response pathway, but for arbitrary response pathways
(e.g., pathways for naming, choice, evaluation, working memory, etc.). In Mozer and Colagrosso (in preparation) we characterize the sufficiency requirements of condition 1; they
involve a representation of low entropy that stays active for long enough that the representation can propagate to the next pathway.
As we will show, a briefly presented stimulus fails to achieve a representation that supports
choice and naming responses. Nonetheless, the stimulus evokes activity in the perceptual
pathway. Because perceptual priming depends on the magnitude of the activation in the
perceptual pathway, not on the activation being communicated to response pathways, the
framework is consistent with the notion of priming occurring in the absence of awareness.
4.1
Simulation of Bar and Biederman (1998)
Bar and Biederman (1998) presented a sequence of masked line drawings of objects and
asked participants to name the objects, even if they had to guess. If the guess was incorrect,
participants were required to choose the object name from a set of four alternatives. Unbeknownst to the participant, some of the drawings in the series were repeated, and Bar and
Biederman were interested in whether participants would benefit from the first presentation even if it could not be identified. The repeated objects could be the same or a different
exemplar of the object, and it could appear in either the same or a different display position.
Participants were able to name 13.5% of drawings on presentation 1, but accuracy jumped
to 34.5% on presentation 2. Accuracy did improve, though not as much, if the same shape
was presented in a different position, but not if a different drawing of the same object
was presented, suggesting a locus of priming early in the visual stream. The improvement
in accuracy is not due to practice in general, because accuracy rose only 4.0% for novel
control objects over the course of the experiment. The priming is firmly subliminal, because
participants were not only unable to name objects on the first presentation, but their fouralternative forced choice (4AFC) performance was not much above chance (28.5%).
To model these phenomena, we created a response pathway with fifty states representing
names of objects that are used in the experiment, e.g., chair and lamp. We also created a
perceptual pathway with states representing visual patterns that correspond to the names
in the response pathway. Following the experimental design, every object identity was
instantiated in two distinct shapes, and every shape could be in one of nine different visualfield positions, leading to 900 distinct states in the perceptual pathway to model the possible
visual stimuli. The following parameters were fit to the data. If two perceptual states, xi
and xk are the same shape in different positions, they are assigned a similarity coefficient
?ik = 0.95; all other similarity coefficients are zero. The association frequency, ?, for
valid associations in the perceptual pathway was 22, and the response pathway 18. Other
parameters were ? p = .05, ? r = .01, and ? = 1.0.
The PIT model achieves a good fit to the human experimental data (Figure 2). Specifically, priming is greatest for the same shape in the same position, some priming occurs for
the same shape in a different position, and no substantial priming occurs for the different
shape. Figure 3a shows the time course of activation of a stimulus representation in the
perceptual pathway when the stimulus is presented for 50 iterations, on both the first and
third presentations. The third presentation was chosen instead of the second to make the
effect of priming clearer.
Even though a shape cannot be named on the first presentation, partial information about
the shape may nonetheless be available for report. The 4AFC test of Bar and Biederman provides a more sensitive measure of residual stimulus information. In past work,
we modeled forced-choice tasks using a response pathway with only the alternatives under
consideration. However, in this experiment, forced-choice performance must be estimated
conditional on incorrect naming. In PIT framework, we achieve this using naming and
40
40
First Block
Second Block
30
25
20
15
10
5
First Block
35
Percent Correct Naming
Percent Correct Naming
35
Second Block
30
25
20
15
10
5
0
0
Control
Objects
Prime
SHAPE: Same
Objects POSITION: Same
Same
Different
Different Different Second
Same
Different Control
Control
Objects
Prime
SHAPE: Same
Objects POSITION: Same
Same
Different
Different Different Second
Same
Different Control
Figure 2: (left panel) Data from Bar and Biederman (1998) (right panel) Simulation of PIT. White
bar: accuracy on first presentation of a prime object. Black bars: the accuracy when the object is
repeated, either with the same or different shape, and in the same or different position. Grey bars:
accuracy for control objects at the beginning and the end of the experiment.
forced-choice output pathways having output distributions N (t) and F (t), which are linked
via the perceptual state, Y p (t). F (t) must be reestimated with the evidence that N (t) is
not the target state. This inference problem is intractable. We therefore used a shortcut in
which a single response pathway is used, augmented with a simple three-node belief net
(Figure 3b) to capture the dependence between naming and forced choice. The belief net
has a response pathway node Y r (t) connected to F (t) and N (t), with conditional distribution P (N (t) = ni |Y r (t) = yj ) = ??ij + (1 ? ?)/|Y r |, and an analogous distribution for
P (F (t) = fi |Y r (t) = yj ). The free parameter ? determines how veridically naming and
forced-choice actions reflect response-pathway output. Over a range of ?, ? < 1, the model
obtains forced-choice performance near chance on the first presentation when the naming
response is incorrect. For example, with ? = 0.72, the model produces a forced-choice
accuracy on presentation 1 of 26.1%. (Interestingly, the model also produces below chance
performance on presentation 2 if the object is not named correctly?23.5%?which is also
found in the human data?20.0%.) Thus, by the stringent criterion of 4AFC, the model
shows no access consciousness, and therefore illustrates a dissociation between priming
and access consciousness. In our simulation, we followed the procedure of Bar and Biederman by including distractor alternatives with visual and semantic similarity to the target.
These distractors are critical: with unrelated distractors, the model?s 4AFC performance is
significantly above chance, illustrating that a perceptual representation can be adequate to
support some responses but not others, as Farah et al. (1994) also argued.
4.2
Simulation of Abrams and Greenwald (2000)
During an initial phase of the experiment, participants categorized 24 clearly visible target
words as pleasant (e.g., HUMOR) or unpleasant (e.g., SMUT). They became quite familiar
with the task by categorizing each word a total of eight times. In a second phase, participants were asked to classify the same targets and were given a response deadline to induce
errors. The targets were preceded by masked primes that could not be identified. Of interest
is the effective valence (or EV) of the target for different prime types, defined as the error
rate difference between unpleasant and pleasant targets. A positive (negative) EV indicates
that responses are biased toward a pleasant (unpleasant) interpretation by the prime. As one
would expect, pleasant primes resulted in a positive EV, unpleasant primes in a negative EV.
Of critical interest is the finding that a nonword prime formed by recombining two pleasant targets (e.g., HULIP from HUMOR and TULIP) or unpleasant targets (e.g., BIUT from
BILE and SMUT ) also served to bias the targets. More surprising, a positive EV resulted
from unpleasant prime words formed by recombining two pleasant targets (TUMOR from
TULIP and HUMOR ), indicating that subliminal priming arises from word fragments, not
words as unitary entities, and providing further evidence for an early locus of subliminal
priming. Note that the results depend critically on the first phase of the experiment, which
gave participants extensive practice on a relatively small set of words that were then used
as and recombined to form primes. Words not studied in the first phase (orphans) provided
Probability
0.6
0.5
0.4
0.3
0.2
0.1
0
object, first presentation
object, third presentation
different object
N(t)
F(t)
Yr(t)
1
50
1000
(a)
(b)
Time (msec)
Figure 3: (a) Activation of the perceptual representation in PIT as a function of processing iterations
Effective Valence
on the ?rst (thin solid line) and third (thick solid line) presentations of target. (b) Bayes net for
performing 4AFC conditional on incorrect naming response.
0.4
Experiment
Model
0.3
0.2
0.1
0
targets
hulip-type
tumor-type
orphans
Figure 4: Effective valence of primes in the Abrams and Greenwald (2000) experiment for human
subjects (black bars) and PIT model (grey bars). HULIP-type primes are almost as strong as target
repetitions, and TUMOR-type primes have a positive valence, contrary to the meaning of the word.
no signi?cant EV effect when used as primes.
In this simulation, we used a three pathway model: a perceptual pathway that maps visual patterns to orthography with 200 input states corresponding both to words, nonwords,
and nonword recombinations of words; a semantic pathway that maps to 100 distinct lexical/semantic states; and a judgement pathway that maps to two responses, pleasant and
unpleasant. In the perceptual pathway, similarity structure was based on letter overlap,
so that HULIP was similar to both TULIP and HUMOR, with ? = 0.837. No similarity
was assumed in the semantic state representation; consistent with the previous simulation,
? p = .05, ? s = .01, ? j = .01, and ? = .01. At the outset of the simulation, ? frequencies
for correct associations were 15, 19, and 25 in the perceptual, semantic, and judgement
pathways. The initial phase of the experiment was simulated by repeated supraliminal
presentation of words, which increased the association frequencies in all three pathways
through the ??ij learning rule.
Long-term supraliminal priming is essential in establishing the association strengths, as
we?ll explain. Short-term subliminal priming also plays a key role in the experiment. During the second phase of the experiment, residual activity from the prime?primarily in the
judgement pathway?biases the response to the target. Residual activation of the prime
is present even if the representation of the prime does not reach suf?cient strength that it
could be named or otherwise reported.
The outcome of the simulation is consistent with the human data (Figure 4). When a
HULIP -type prime is presented, HUMOR and TULIP become active in the semantic pathway because of their visual similarity to HULIP. Partial activation of these two practiced
words pushes the judgement pathway toward a pleasant response, resulting in a positive
EV. When a TUMOR-type prime is presented, three different words become active in the
semantic pathway: HUMOR, TULIP, and TUMOR itself. Although TUMOR is more active, it
was not one of the words studied during the initial phase of the experiment, and as a result,
it has a relatively weak association to the unpleasant judgement, in contrast to the other
two words which have strong associations to the pleasant judgement. Orphan primes have
little effect because they were not studied during the initial phase of the experiment, and
consequently their association to pleasant and unpleasant judgements is also weak. In summary, activation of the prime along a critical, well-practiced pathway may not be suf?cient
to support an overt naming response, yet it may be suf?cient to bias the processing of the
immediately following target.
5
Discussion
An important contribution of this work has been to demonstrate that specific experimental
results relating to access consciousness and subliminal priming can be interpreted in a concrete computational framework. By necessity, the PIT framework, which we previously
used to model supraliminal priming data, predicts the existence of subliminal priming,
because the mechanisms giving rise to priming depend on degree of activation of a representation, whereas the processes giving rise to access consciousness also depend on the
temporal persistence of a representation.
Another contribution of this work has been to argue that two previous computational models each tell only part of the story. Farah et al. argue that quality of representation is
critical; Dehaene et al. argue that pathways to communicate representations is critical. The
PIT framework argues that both of these features are necessary for access consciousness.
Although the PIT framework is not completely developed, it nonetheless makes a clear
prediction: that subliminal priming is can never be stronger than supraliminal priming,
because the maximal activation of subliminal primes is never greater than that of supraliminal primes. One might argue that many theoretical frameworks might predict the same,
but no other computational model is sufficiently well developed?in terms of addressing
both priming and access consciousness?to make this prediction.
In its current stage of development, a weakness of the PIT framework is that it is silent
as to how perceptual and response pathways become flexibly interconnected based on task
demands. However, the PIT framework is not alone in failing to address this critical issue:
The Dehaene et al. model suggests that once a representation enters the global workspace,
all response modules can access it, but the model does not specify how the appropriate perceptual module wins the competition to enter the global workspace, or how the appropriate
response module is activated. Clearly, flexible cognitive control structures that perform
these functions are intricately related to mechanisms of consciousness.
Acknowledgments
This research was supported by NIH/IFOPAL R01 MH61549?01A1.
References
Abrams, R. L., & Greenwald, A. G. (2000). Parts outweigh the whole (word) in unconscious analysis
of meaning. Psychological Science, 11(2), 118?124.
Baars, B. (1989). A cognitive theory of consciousness. Cambridge: Cambridge University Press.
Bar, M., & Biederman, I. (1998). Subliminal visual priming. Psychological Science, 9(6), 464?468.
Block, N. (1995). On a confusion about a function of consciousness. Brain and Behavioral Sciences,
18(2), 227?247.
Dehaene, S., & Naccache, L. (2001). Towards a cognitive neuroscience of consciousness: basic
evidence and a workspace framework. Cognition, 79, 1?37.
Dehaene, S., Sergent, C., & Changeux, J.-P. (2003). A neuronal network model linking subjective reports and objective physiological data during conscious perception. Proceedings of the
National Academy of Sciences, 100, 8520?8525.
Farah, M. J., O?Reilly, R. C., & Vecera, S. P. (1994). Dissociated overt and covert recognition as an
emergent property of a lesioned neural network. Psychological Review, 100, 571?588.
Mathis, D. W., & Mozer, M. C. (1996). Conscious and unconscious perception: a computational
theory. In G. Cottrell (Ed.), Proceedings of the Eighteenth Annual Conference of the Cognitive
Science Society (pp. 324?328). Hillsdale, NJ: Erlbaum & Associates.
Mozer, M. C., Colagrosso, M. D., & Huber, D. E. (2002). A rational analysis of cognitive control
in a speeded discrimination task. In T. G. Dietterich, S. Becker, & Z. Ghahramani (Eds.),
Advances in Neural Information Processing Systems 14. Cambridge, MA: MIT Press.
Mozer, M. C., Colagrosso, M. D., & Huber, D. E. (2003). Mechanisms of long-term repetition
priming and skill refinement: A probabilistic pathway model. In Proceedings of the TwentyFifth Annual Conference of the Cognitive Science Society. Hillsdale, NJ: Erlbaum Associates.
| 2715 |@word trial:3 illustrating:1 briefly:1 version:1 judgement:7 seems:1 stronger:1 grey:2 simulation:10 propagate:1 solid:4 necessity:1 initial:4 series:4 fragment:1 practiced:2 interestingly:1 past:4 reaction:1 subjective:1 current:1 surprising:1 activation:14 yet:2 must:7 skepticism:1 subsequent:1 visible:1 cottrell:1 cant:1 shape:13 drop:1 exploded:1 discrimination:1 alone:1 guess:2 item:2 yr:1 xk:2 beginning:1 lamp:1 short:1 mental:2 provides:2 coarse:1 node:2 successive:1 along:1 become:4 ik:3 incorrect:4 consists:1 poised:3 pathway:63 behavioral:1 manner:1 huber:5 rapid:1 embody:1 behavior:4 distractor:1 brain:4 little:1 delineation:1 considering:1 increasing:1 revision:1 becomes:1 provided:1 underlying:1 unrelated:1 panel:2 what:2 interpreted:2 developed:2 finding:1 nj:2 temporal:2 y3:2 every:2 golden:1 act:1 veridically:1 control:10 hillary:1 appear:1 t1:7 positive:5 influencing:1 despite:1 accumulates:1 chocolate:1 establishing:1 might:3 black:2 studied:3 suggests:2 pit:15 co:2 range:1 tulip:5 speeded:1 acknowledgment:1 yj:8 practice:3 block:7 implement:1 communicated:3 x3:2 procedure:1 area:4 cascade:1 printed:1 reilly:3 persistence:2 word:21 confidence:1 refers:2 significantly:1 induce:1 outset:1 cannot:2 clever:1 prosopagnosia:2 storage:1 influence:2 bill:1 map:5 outweigh:1 lexical:1 eighteenth:1 exposure:1 attention:1 flexibly:1 formalized:1 immediately:1 rule:1 facilitation:1 handle:1 notion:4 analogous:1 unconscious:4 target:19 colorado:3 damaged:1 strengthen:1 exact:1 play:2 trigger:1 associate:2 recognition:4 particularly:1 updating:1 predicts:1 observed:1 role:1 module:7 enters:1 capture:2 parameterize:1 connected:1 contemporary:1 nonwords:1 rose:1 mozer:13 environment:1 substantial:1 asked:4 lesioned:1 mine:2 dynamic:2 depend:3 serve:1 efficiency:3 basis:1 completely:1 emergent:1 alphabet:1 distinct:7 forced:9 describe:4 instantiated:1 effective:3 pertaining:1 detected:1 tell:1 outcome:1 aloud:1 whose:2 lag:1 quite:2 plausible:1 say:2 drawing:4 interconnection:1 otherwise:1 ability:3 itself:1 sequence:3 net:4 propose:1 interconnected:3 maximal:1 causing:1 philosophy:1 rapidly:1 promiscuous:1 achieve:3 academy:1 competition:1 rst:1 transmission:6 speedaccuracy:1 requirement:1 produce:4 object:21 tions:1 weakens:1 clearer:1 exemplar:1 ij:7 school:1 strong:2 implemented:1 involves:1 indicate:1 signi:1 met:1 thick:1 correct:4 human:4 stringent:1 hillsdale:2 explains:1 require:1 premise:1 argued:1 formedness:1 recombined:1 sufficiently:3 residing:1 visually:2 great:1 cognition:4 predict:1 claim:1 jumped:1 achieves:1 early:2 purpose:1 perceived:2 failing:1 overt:3 xnx:1 coordination:1 sensitive:1 repetition:6 successfully:1 mit:1 clearly:2 primed:1 categorizing:1 focus:3 improvement:2 indicates:2 likelihood:1 baars:2 contrast:2 mh61549:1 inference:3 interconnect:1 hidden:3 interested:1 issue:3 fidelity:1 among:1 flexible:1 denoted:2 development:2 resonance:2 constrained:1 once:1 never:2 having:2 afc:5 thin:1 future:1 others:3 report:6 stimulus:10 t2:6 inherent:1 few:2 primarily:1 resulted:2 national:1 familiar:2 phase:10 fire:1 attractor:1 maintain:1 attempt:2 interest:2 highly:1 evaluation:3 weakness:1 yielding:1 activated:1 explosion:2 partial:2 experience:5 necessary:1 shorter:1 phenomenal:2 theoretical:2 minimal:2 psychological:4 increased:4 classify:1 modeling:1 localist:1 strategic:1 addressing:2 rare:1 hundred:1 masked:2 erlbaum:2 characterize:2 reported:2 supposes:1 fundamental:2 sensitivity:1 stay:1 workspace:9 probabilistic:5 intricately:1 michael:2 reestimated:1 concrete:3 connectivity:2 again:1 central:1 reflect:1 choose:1 cognitive:9 inferentially:1 leading:1 colagrosso:7 account:6 suggesting:1 ifopal:1 availability:1 includes:1 coefficient:2 coordinated:1 depends:1 stream:1 linked:1 characterizes:3 bayes:2 participant:13 complicated:1 farah:9 slope:1 contribution:3 formed:2 ni:1 accuracy:12 became:1 characteristic:1 dissociation:2 correspond:1 identify:2 blink:2 consciously:2 famous:3 identification:2 bayesian:1 weak:2 critically:1 served:1 explain:3 reach:1 ed:2 yny:1 nonetheless:6 frequency:6 pp:1 proof:1 demystify:1 associated:1 rational:2 realm:1 knowledge:1 distractors:2 appears:2 higher:1 reflected:1 response:35 specify:2 sufficiency:1 though:3 just:1 stage:1 working:2 quality:4 perhaps:1 scientific:2 name:11 usa:2 effect:4 requiring:2 y2:2 usage:1 dietterich:1 assigned:1 read:1 entering:1 consciousness:29 semantic:8 white:1 ll:1 during:7 criterion:1 presenting:1 outline:1 demonstrate:1 confusion:1 covert:2 argues:1 bring:1 percent:2 reasoning:1 image:1 meaning:2 consideration:1 novel:2 recently:1 fi:1 nih:1 common:1 speculative:1 specialized:1 functional:2 preceded:2 discriminated:1 association:15 interpretation:1 linking:1 relating:1 functionally:1 significant:2 refer:1 theorist:1 cambridge:3 enter:2 similarly:1 had:1 access:21 similarity:9 etc:1 posterior:2 recent:1 perspective:1 prime:26 certain:2 yi:3 greater:3 converge:1 paradigm:1 relates:1 multiple:3 hebbian:1 faster:1 long:6 concerning:2 naming:14 equally:1 deadline:1 a1:1 prediction:2 basic:2 patient:2 nonword:2 iteration:3 orthography:1 achieved:4 condi:1 whereas:3 addressed:1 wealth:1 fifty:1 biased:1 subject:1 dehaene:10 contrary:1 incorporates:1 habituation:1 seem:1 unitary:1 near:1 enough:1 variety:2 fit:2 gave:1 architecture:5 identified:3 restrict:1 silent:1 shift:1 whether:1 six:1 becker:1 effort:3 speech:1 nine:1 action:2 adequate:1 clear:2 involve:2 pleasant:10 conscious:9 specifies:1 exist:1 delta:1 estimated:1 correctly:1 neuroscience:1 discrete:2 key:1 four:1 diffusion:2 utilize:1 imaging:1 year:1 letter:9 communicate:2 named:4 evokes:1 almost:1 reader:1 decision:1 incompatible:1 followed:1 display:1 annual:2 activity:2 strength:5 kronecker:1 x2:2 aspect:3 speed:2 chair:1 performing:1 recombining:2 relatively:2 department:1 describes:1 y0:1 evolves:1 lasting:1 gradually:1 boulder:1 unbeknownst:1 previously:1 pin:1 mechanism:7 locus:2 mechanistic:1 serf:3 end:1 available:3 operation:1 eight:1 appropriate:2 alternative:7 shortly:2 existence:2 embodies:1 giving:2 recombination:1 ghahramani:1 classical:1 r01:1 society:2 objective:1 question:1 occurs:3 damage:1 dependence:1 traditional:1 exhibit:1 abrams:3 win:1 distance:1 deficit:1 attentional:2 unable:1 sergent:3 accessibility:1 hmm:2 nx:1 valence:4 entity:1 naccache:2 argue:8 simulated:1 reason:1 toward:3 modeled:3 relationship:1 copying:1 providing:2 potentially:1 negative:2 rise:2 ordinarily:1 suppress:1 design:2 perform:2 upper:3 neuron:1 observation:1 markov:2 y1:2 distracted:1 arbitrary:2 biederman:7 inferred:1 required:2 extensive:1 connection:1 philosophical:1 address:5 able:1 bar:12 below:1 perception:3 pattern:4 ev:7 summarize:1 built:1 reliable:1 memory:3 including:1 belief:3 greatest:1 critical:8 overlap:1 cascaded:1 residual:4 representing:2 improve:1 x2i:2 firmly:1 picture:1 created:2 deemed:1 carried:1 kj:1 prior:4 literature:1 understanding:3 review:1 catalyst:1 expect:1 suf:3 limitation:2 integrate:1 awareness:2 ripe:1 degree:1 sufficient:1 consistent:3 propagates:1 story:1 maxt:1 vecera:3 course:7 summary:1 supported:1 free:1 verbal:2 bias:3 allow:2 institute:1 face:5 distributed:2 benefit:1 feedback:1 curve:3 cortical:1 transition:2 valid:1 made:1 refinement:2 qualitatively:1 collection:1 far:1 skill:2 obtains:1 overcomes:1 global:5 active:4 assumed:1 xi:9 nature:1 orphan:3 dissociated:1 priming:40 mathis:4 complex:1 domain:2 clinton:1 assured:1 did:1 whole:1 repeated:4 categorized:1 x1:2 augmented:1 neuronal:1 cient:3 elaborate:1 ny:1 fails:1 position:9 msec:1 perceptual:27 answering:1 third:6 minute:1 familiarity:1 specific:5 changeux:2 showing:1 subliminal:15 list:2 decay:1 physiological:1 evidence:4 essential:2 intractable:1 magnitude:1 illustrates:2 occurring:2 push:1 demand:1 entropy:1 photograph:1 simply:3 explore:1 likely:1 visual:10 loses:1 chance:4 determines:1 ma:1 conditional:3 identity:5 presentation:15 greenwald:3 consequently:1 towards:1 absence:3 content:1 change:1 shortcut:1 typical:2 determined:1 specifically:1 semantically:1 tumor:6 total:1 experimental:8 indicating:1 unpleasant:9 internal:2 support:3 arises:1 categorize:2 frontal:3 preparation:1 phenomenon:8 |
1,889 | 2,716 | Bayesian inference in spiking neurons
Sophie Deneve?
Gatsby Computational Neuroscience Unit
University College London
London, UK WC1N 3AR
[email protected]
Abstract
We propose a new interpretation of spiking neurons as Bayesian integrators accumulating evidence over time about events in the external world
or the body, and communicating to other neurons their certainties about
these events. In this model, spikes signal the occurrence of new information, i.e. what cannot be predicted from the past activity. As a result,
firing statistics are close to Poisson, albeit providing a deterministic representation of probabilities. We proceed to develop a theory of Bayesian
inference in spiking neural networks, recurrent interactions implementing a variant of belief propagation.
Many perceptual and motor tasks performed by the central nervous system are probabilistic, and can be described in a Bayesian framework [4, 3]. A few important but hidden
properties, such as direction of motion, or appropriate motor commands, are inferred from
many noisy, local and ambiguous sensory cues. These evidences are combined with priors
about the sensory world and body. Importantly, because most of these inferences should
lead to quick and irreversible decisions in a perpetually changing world, noisy cues have to
be integrated on-line, but in a way that takes into account unpredictable events, such as a
sudden change in motion direction or the appearance of a new stimulus.
This raises the question of how this temporal integration can be performed at the neural
level. It has been proposed that single neurons in sensory cortices represent and compute
the log probability that a sensory variable takes on a certain value (eg Is visual motion in
the neuron?s preferred direction?) [9, 7]. Alternatively, to avoid normalization issues and
provide an appropriate signal for decision making, neurons could represent the log probability ratio of a particular hypothesis (eg is motion more likely to be towards the right than
towards the left) [7, 6]. Log probabilities are convenient here, since under some assumptions, independent noisy cues simply combine linearly. Moreover, there are physiological
evidence for the neural representation of log probabilities and log probability ratios [9, 6, 7].
However, these models assume that neurons represent probabilities in their firing rates. We
argue that it is important to study how probabilistic information are encoded in spikes.
Indeed, it seems spurious to marry the idea of an exquisite on-line integration of noisy cues
with an underlying rate code that requires averaging on large populations of noisy neurons
and long periods of time. In particular, most natural tasks require this integration to take
place on the time scale of inter-spike intervals. Spikes are more efficiently signaling events
?
Institute of Cognitive Science, 69645 Bron, France
than analog quantities. In addition, a neural theory of inference with spikes will bring us
closer to the physiological level and generate more easily testable predictions.
Thus, we propose a new theory of neural processing in which spike trains provide a deterministic, online representation of a log-probability ratio. Spikes signals events, eg that
the log-probability ratio has exceeded what could be predicted from previous spikes. This
form of coding was loosely inspired by the idea of ?energy landscape? coding proposed
by Hinton and Brown [2]. However, contrary to [2] and other theories using rate-based
representation of probabilities, this model is self-consistent and does not require different
models for encoding and decoding: As output spikes provide new, unpredictable, temporally independent evidence, they can be used directly as an input to other Bayesian neurons.
Finally, we show that these neurons can be used as building blocks in a theory of approximate Bayesian inference in recurrent spiking networks. Connections between neurons
implement an underlying Bayesian network, consisting of coupled hidden Markov models.
Propagation of spikes is a form of belief propagation in this underlying graphical model.
Our theory provides computational explanations of some general physiological properties
of cortical neurons, such as spike frequency adaptation, Poisson statistics of spike trains,
the existence of strong local inhibition in cortical columns, and the maintenance of a tight
balance between excitation and inhibition. Finally, we discuss the implications of this
model for the debate about temporal versus rate-based neural coding.
1 Spikes and log posterior odds
1.1 Synaptic integration seen as inference in a hidden Markov chain
We propose that each neuron codes for an underlying ?hidden? binary variable, xt , whose
state evolves over time. We assume that xt depends only on the state at the previous time
step, xt?dt , and is conditionally independent of other past states. The state xt can switch
1
limdt?0 P (xt = 1|xt?dt = 0), and from 1 to
from 0 to 1 with a constant rate ron = dt
0 with a constant rate roff . For example, these transition rates could represent how often
motion in a preferred direction appears the receptive field and how long it is likely to stay
there.
The neuron infers the state of its hidden variable from N noisy synaptic inputs, considered
to be observations of the hidden state. In this initial version of the model, we assume
that these inputs are conditionally independent homogeneous Poisson processes, synapse
i
i emitting a spike between time t and t + dt (sit = 1) with constant probability qon
dt if
i
xt = 1, and another constant probability qoff dt if xt = 0. The synaptic spikes are assumed
to be otherwise independent of previous synaptic spikes, previous states and spikes at other
synapses. The resulting generative model is a hidden Markov chain (figure 1-A).
However, rather than estimating the state of its hidden variable and communicating this
estimate to other neurons (for example by emitting a spike when sensory evidence for
xt = 1 goes above a threshold) the neuron reports and communicates its certainty that the
current state is 1. This certainty takes the form of the log of the ratio of the probability
that the hidden state is 1, and the probability that the state is 0, given all the synaptic inputs
? P (xt =1|s0?t ) ?
received so far: Lt = log P
(xt =0|s0?t ) . We use s0?t as a short hand notation for the N
synaptic inputs received at present and in the past. We will refer to it as the log odds ratio.
Thanks to the conditional independencies assumed in the generative model, we can compute this Log odds ratio iteratively. Taking the limit as dt goes to zero, we get the following
differential equation:
?
?
?
? P
L? = ron 1 + e?L ? roff 1 + eL + i wi ?(sit ? 1) ? ?
B.
A.
xt dt
ron .roff
qon , qoff
st dt
xt
ron .roff
xt dt
qon , qoff
st
i
t
s
qon , qoff
st
Ot
j
It
?
Ot
Lt
st dt
?
Gt
t
t
C.
E.
2
0
-2
-4
D.
500
1000
1500
2000
2500
2
3000
Count
Log odds
4
20
Lt 0
-2
0
500
1000
1500
2000
2500
Time
Ot
3000
0
200
400
600
ISI
Figure 1: A. Generative model for the synaptic input. B. Schematic representation of log
odds ratio encoding and decoding. The dashed circle represents both eventual downstream
elements and the self-prediction taking place inside the model neuron. A spike is fired only
when Lt exceeds Gt . C. One example trial, where the state switches from 0 to 1 (shaded
area) and back to 0. plain: Lt , dotted: Gt . Black stripes at the top: corresponding spikes
train. D. Mean Log odds ratio (dark line) and mean output firing rate (clear line). E. Output
spike raster plot (1 line per trial) and ISI distribution for the neuron shown is C. and D.
Clear line: ISI distribution for a poisson neuron with the same rate.
wi , the synaptic weight, describe how informative synapse i is about the state of the hidden
? qi ?
variable, e.g. wi = log qion . Each synaptic spike (sit = 1) gives an impulse to the log
off
odds ratio, which is positive if this synapse is more active when the hidden state if 1 (i.e it
increases the neuron?s confidence that the state is 1), and negative if this synapse is more
active when xt = 0 (i.e it decreases the neuron?s confidence that the state is 1).
The bias, ?, is determined by how informative it is not to receive any spike, e.g. ? =
P
i
i
i qon ? qoff . By convention, we will consider that the ?bias? is positive or zero (if not,
we need simply to invert the status of the state x).
1.2 Generation of output spikes
The spike train should convey a sparse representation of Lt , so that each spike reports new
information about the state xt that is not redundant with that reported by other, preceding,
spikes. This proposition is based on three arguments: First, spikes, being metabolically
expensive, should be kept to a minimum. Second, spikes conveying redundant information
would require a decoding of the entire spike train, whereas independent spike can be taken
into account individually. And finally, we seek a self consistent model, with the spiking
output having a similar semantics to its spiking input.
To maximize the independence of the spikes (conditioned on xt ), we propose that the neuron fires only when the difference between its log odds ratio Lt and a prediction Gt of this
log odds ratio based on the output spikes emitted so far reaches a certain threshold. Indeed,
supposing that downstream elements predicts Lt as best as they can, the neuron only needs
to fire when it expects that prediction to be too inaccurate (figure 1-B). In practice, this
will happen when the neuron receives new evidence for xt = 1. Gt should thereby follow
the same dynamics as Lt when spikes are not received. The equation for Gt and the output
Ot (Ot = 1 when an output spike is fired) are given by:
G? =
Ot
=
?
?
?
?
ron 1 + e?L ? roff 1 + eL + go ?(Ot ? 1)
go
1. when Lt > Gt + , 0 otherwise,
2
(1)
(2)
Here go , a positive constant, is the only free parameter, the other parameters being constrained by the statistics of the synaptic input.
1.3 Results
Figure 1-C plots a typical trial, showing the behavior of L, G and O before, during and after
presentation of the stimulus. As random synaptic inputs are integrated, L fluctuates and
eventually exceeds G + 0.5, leading to an output spike. Immediately after a spike, G jumps
to G + go , which prevents (except in very rare cases) a second spike from immediately
following the first. Thus, this ?jump? implements a relative refractory period.
? on ? However,
. Thus L
G decays as it tends to converge back to its stable level gstable = log rroff
eventually exceeds G again, leading to a new spike. This threshold crossing happens more
often during stimulation (xt = 1) as the net synaptic input alters to create a higher overall
level of certainty, Lt .
Mean Log odds ratio and output firing rate
? t of the Bayesian neuron during presentation of its preferred stimulus
The mean firing rate O
(i.e. when xt switches from 0 to 1 and back to 0) is plotted in figure 1-D, together with the
? t , both averaged over trials. Not surprisingly, the log-posterior
mean log posterior ratio L
ratio reflects the leaky integration of synaptic evidence, with an effective time constant that
depends on the transition probabilities ron , roff . If the state is very stable (ron = roff ? 0),
synaptic evidence is integrated over almost infinite time periods, the mean log posterior
ratio tending to either increase or decrease linearly with time. In the example in figure 1D, the state is less stable, so ?old? synaptic evidence are discounted and Lt saturates.
? t tracks the state of xt almost perfectly. This
In contrast, the mean output firing rate O
is because, as a P
form of predictive coding, the output spikes reflect the new synaptic
i
evidence, It =
i ?(st ? 1) ? ?, rather than the log posterior ratio itself. In particular, the mean output firing rate is a rectified linear function of the mean input, e. g.
h
i+
? = 1 I? = P wi q i
?? .
O
go
i
on(off)
Analogy with a leaky integrate and fire neuron
We can get an interesting insight into the computation performed by this neuron by linearizing L and G around their mean levels over trials. Here we reduce the analysis to prolonged,
statistically stable periods when the state is constant (either ON or OFF). In this case, the
? and its output prediction G
? are also constant over time. We make
mean level of certainty L
the rough approximation that the post spike jump, go , and the input fluctuations are small
?
compared to the mean level of certainty L.
Rewriting Vt = Lt ? Gt +
go
2
as the ?membrane potential? of the Bayesian neuron:
V? = ?kL? V + It ? ?go ? go Ot
?
?
where kL? = ron e?L + roff eL , the ?leak? of the membrane potential, depends on the overall
level of certainty. ?go is positive and a monotonic increasing function of go .
A.
s t1 dt
s t1
s t1 dt
B.
C.
x t1
x t3 dt
x t3
x t3 dt
x t1 dt
x t1
x t1 dt
x t2
x t3
x t1
?
x tn
x t3
x t2
?
x tn
?
Lx2
D.
x t2 dt
s t2 dt
x t2
s t2
x t2 dt
s t2 dt
Log odds
10
No inh
-0.5
-1
-1
-2-1.5
5
Feedback
500 1000 1500 2000
Tiger
Stripes
0
-5
-10
500
1000
1500
2000
2500
Time
Figure 2: A. Bayesian causal network for yt (tiger), x1t (stripes) and x2t (paws). B. A network feedforward computing the log posterior for x1t . C. A recurrent network computing
the log posterior odds for all variables. D. Log odds ratio in a simulated trial with the net2
1
1
work in C (see text). Thick line: Lxt , thin line: Lxt , dash-dotted: Lxt without inhibition.
2
Insert: Lxt averaged over trials, showing the effect of feedback.
The linearized Bayesian neuron thus acts in its stable regime as a leaky integrate and fire
(LIF) neuron. The membrane potential Vt integrates its input, Jt = It ? ?go , with a leak
kL? . The neuron fires when its membrane potential reaches a constant threshold go . After
each spikes, Vt is reset to 0.
Interestingly, for appropriately chosen compression factor go , the mean input to the linearized neuron J? = I? ? ?go ? 0 1 . This means that the membrane potential is purely
driven to its threshold by input fluctuations, or a random walk in membrane potential. As
a consequence, the neuron?s firing will be memoryless, and close to a Poisson process. In
particular, we found Fano factor close to 1 and quasi-exponential ISI distribution (figure 1E) on the entire range of parameters tested. Indeed, LIF neurons with balanced inputs have
been proposed as a model to reproduce the statistics of real cortical neurons [8]. This balance is implemented in our model by the neuron?s effective self-inhibition, even when the
synaptic input itself is not balanced.
Decoding
As we previously said, downstream elements could predict the log odds ratio Lt by computing Gt from the output spikes (Eq 1, fig 1-B). Of course, this requires an estimate of
the transition probabilities ron , roff , that could be learned from the observed spike trains.
However, we show next that explicit decoding is not necessary to perform bayesian inference in spiking networks. Intuitively, this is because the quantity that our model neurons
receive and transmit, eg new information, is exactly what probabilistic inference algorithm
propagate between connected statistical elements.
1
Even if go is not chosen optimally, the influence of the drift J? is usually negligible compared to
the large fluctuations in membrane potential.
2 Bayesian inference in cortical networks
The model neurons, having the same input and output semantics, can be used as building blocks to implement more complex generative models consisting of coupled Markov
chains. Consider, for example, the example in figure 2-A. Here, a ?parent? variable x1t
(the presence of a tiger) can cause the state of n other ?children? variables ([xkt ]k=2...n ),
of whom two are represented (the presence of stripes,x2t , and motion, x3t ). The ?children? variables are Bayesian neurons identical to those described previously. The resulting
bayesian network consist of n + 1 coupled hidden Markov chains. Inference in this architecture corresponds to computing the log posterior odds ratio for the tiger, x1t , and the log
posterior of observing stripes or motion, ([xkt ]k=2...n ), given the synaptic inputs received
by the entire network so far, i.e. s20?t , . . . , sk0?t .
Unfortunately, inference and learning in this network (and in general in coupled Markov
chains) requires very expensive computations, and cannot be performed by simply propagating messages over time and among the variable nodes. In particular, the state of a child
variable xtk depends on xkt?dt , skt , x1t and the state of all other children at the previous
time step, [xjt?dt ]2<j<n,j6=i . In contrast, our network can only implement pairwise interactions, a connection between two spiking neurons implementing the conditional probability
linking the two corresponding binary variables.
Thus, we need to assume additional conditional independencies between the nodes in the
generative model, so that their joint probability can be pairwise factorized: p(xt , xt?1 ) =
Q
j Q
1
i
i
i
ij ?(xt , xt )
i ?(xt , xt?dt ). In words, it means that variables bias each other?s probZ
abilities, but do not influence each other?s dynamics, i.e they do not affect each other?s
transition probabilities. For example, a tiger does not affect the probability that stripes
appear or disappear, but increases their probability of being present.
Naive implementation
In this restricted case, marginal posterior probabilities can be computed iteratively by propagating beliefs in time and between the variables, or, in our model, by propagating spikes
in a neural network. This is because the probability of a variable xkt can be directly updated by the conditional probability
of observing the synaptic input to another connected
P
neuron, slt , eg p(slt |xkt ) = xl p(slt |xlt )p(xlt |xkt ), marginalizing out the hidden state xl . Of
t
course, rather than using slt , we use Otl , the output of the Bayesian neuron coding for xlt .
As we said previously, this output directly represents the new synaptic evidence received
by neuron l. The resulting equation is identical to the one derived previously for poisson
input,
P
k
k
L? x = fk (Lx ) + l wlk ?(Otl ? 1) ? ?k
k
k
(1 + e?x ) ? roff
(1 + ex ). As previously, wlk , the synaptic weight,
where fk (x) = ron
describes how informative it is for neuron k to receive a spike from neuron (or synapse)
P (O k =1|y =1)
l, wlk = log( P (Otk =1|ytt =0) ), while ?k is how informative it is not to receive a spike,
t
P
?k = dt l P (Otl = 1|xkt = 1) ? P (Otl = 1|xkt = 0).
This shows that our model is self-consistent. Except at the first stage of processing (eg in
the retina), all inputs are proposed to come from other Bayesian neurons.
Results
We implemented these update rules in a spiking neural network (figure 2-B) representing
the generative model in figure 2-A, with 100 possible children for x1t . We first consider
the case where there is no feedback connections, meaning that w1k = 0 for all k. In this
case the network computes the probability of a tiger at time t, integrating multiple sensory
cues such as the presence of stripes or motion in the visual scene.
In the example trials plotted in figure 2-D, we fixed the state of x1t and x2t : the tiger and
the stripes are present in the shaded temporal window, and absent outside of it. We then
sample the states of the other children (i.e is there motion or not?) and the corresponding
?observed? synaptic inputs, sk0?tmax , from the generative model. Once this synaptic input
has been generated, it is used as an input to the network in figure 2-B. What is plotted in
the Log odds ratio for the tiger, Lxt 1 and the stripes, Lxt 1 , as a function of time. As we can
see, the stripes receive very noisy synaptic input and can only provide weak evidence that
they are present. However, the tiger neuron is able to combine inputs from its 100 children
and get a much higher certainty.
Unfortunately, bayesian inference in this feedforward network is incomplete: The presence
of a tiger affects the probability of stripes, not only the other way round. To implement
this, we also need feedback connections, w1k . The network with feedforward and feedback
processing fails miserably, given that its activity explode, as illustrated in figure 2-D.
Balanced excitation/inhibition
This failure is due to the presence of loops, whereby a spike from neuron k increases
the certainty of neuron l (and its probability of firing) by wkl , and a spike from
neuron l increases in turn the certainty of neuron k by wlk . These loops result in
spikes reverberating through the network, ad infinitum, without reporting new information, a phenomena akin to loopy belief propagation [10]. To avoid overcounting of
evidence, we thus have to discount the reverberated ?old evidence? from the synaptic input:
P
P
k
k
k
L? x = fk (Lx ) + l wkl ?(Otl ? 1) ? l wkl wlk ?(Ot?dt
? 1) ? ?k .
We implemented this discounting using inhibitory neurons recurrently connected to each
excitatory neuron (figure 2-C). The inhibitory neurons are used to predict the redundant
feedback a bayesian neuron will receive and substract this prediction so that, once again,
only new information are taken into account and communicated to other neurons. Each
excitatory loop is compensated by an inhibitory loop, resulting in a balance between
excitation and inhibition at the level of each neuron within the network.
The result on one trial is plotted in figure 2-D. The ?tiger? log odds ratio is almost indistinguishable from the feedforward case, and is not plotted. The ?stripes? Log odds ratio
increases during presentation of the tiger due to the feedback. In other words, the stripes
neuron can take into account no only its own its own synaptic input, but also the synaptic
input to the other children neurons (such as evidence for motion), thanks to the presence of
a common source (the tiger).
Over many trials, we found that the statistics of the bayesian neuron were still poisson,
and their output firing is stillP
a rectified linear function of the input firing rate in a stable
?l +
?k = [
statistical regime, i.e. O
l wkl O ] .
Discussion
We started from an interpretation of synaptic integration in single neurons as a form of
inference in a hidden Markov chain. We derived a model of spiking neurons and their
interactions able to compute the marginal posterior probabilities of sensory and motor variables given evidence received in the entire network. In this view, the brain implements
an underlying bayesian network in an interconnected neural architecture, with conditional
probabilities represented by synaptic weights. The model makes a rich set of predictions
for the general properties of neuron and synaptic dynamics, such as a time constant that
depends on the overall level of inputs, specific forms of frequency dependant spike and
synaptic adaptation (not shown here) and micro-balanced excitation and inhibition. However, it is still restricted to probabilistic computations involving binary variables. In a
related work similar ideas are applied population encoding of log probability distribution
for analog variables (Zemel, Huys and Dayan, submitted to NIPS 2004).
Despite non-linear processing at the single neural level, the emerging picture is relatively
simple: The neuron acts as a leaky integrate and fire neuron driven by noise. The output
firing rate is a rectified weighted sum of the input firing rates, while the firing statistics are
Poisson. However, these output spike trains are a deterministic function of the input spike
trains. Spikes report fluctuations in the level of certainty that could not be predicted either
from the stability of its stimulus (contribution from Gt ) or the loops in the network (contribution from the inhibitory neuron). Thus firing will be, by definition, unpredictable. This
last observation leads us to suggest that the irregular firing and Poisson statistics observed
in cortical neurons [1] arises as a direct consequence of the random fluctuations in the sensory inputs and the instability of the real word, but are not due to unreliable or ?chaotic?
neural processing.
Finally, it is crucial for the biological realism of the model to find adaptive neural dynamics
and synaptic plasticity able to learn the parameters of the internal model and conditional
probabilities, and we are currently exploring these issues. Fortunately, the required learning
rules are local and unsupervised. According to our preliminary work, the synaptic weights
and bias depend on the joint probability of presynaptic/postsynaptic spikes and can be
learned with the spike time dependent plasticity observed in hippocampus and cortex [5].
Meanwhile, the transition probabilities simply correspond to how often the neuron switches
between an active and an inactive state.
Acknowledgments
We thank Peter Dayan, Peter Latham, Zoubin Ghahramani, Jean Laurens and Jacques
Droulez for helpful discussions and comments. This work was supported by a Dorothy
Hodgkin fellowship for the Royal Society and the BIBA European Project.
References
[1] K. H. Britten, M. N. Shadlen, W. T. Newsome, and J. A. Movshon. The analysis of visual
motion: A comparison of neuronal and psychophysical performance. Journal of Neuroscience,
12:4745?4765, 1992.
[2] G. Hinton and A. Brown. Spiking boltzmann machines. In S. Solla, T. Leen, and K. Muller, editors, Neural Information Processing System, volume 12, pages 122?8. MIT Press, Cambridge,
MA, 2000.
[3] D. Knill and W. Richards. Perception as Bayesian inference. Cambridge University Press,
Cambridge, MA, 1996.
[4] K. Kording and D. Wolpert. Bayesian integration in sensorimotor learning. Nature, 427:244?7,
2004.
[5] H. Markram and M. Tsodyks. Redistribution of synaptic efficacy between neocortical pyramidal
neurons. Nature, 382:807?19, 1996.
[6] M. Mazurek, J. Roitman, J. Ditterich, and M. Shadlen. A role for neural integrators in perceptual
decision making. Cerebral cortex, 13(11):1257?69, 2003.
[7] R. Rao. Bayesian computation in recurrent neural circuits. Neural Computation, 16(1):1?38,
2003.
[8] M. Shadlen and W. Newsome. Noise, neural codes and cortical organization. Current Opinion
in Neurobiology, 4:569?579, 1994.
[9] Y. Weiss and D. Fleet. Velocity likelihood in biological and machine vision. In R. Rao, B. Olshausen, and M. Lewicki, editors, Probabilistic Models of the Brain: Perception and Neural
Function, pages 77?96. MIT Press, Cambridge, MA, 2002.
[10] Y. Weiss and W. Freeman. Correctness of belief propagation in gaussian graphical models of
arbitrary topology. Neural Computation, 13:2173?200, 2001.
| 2716 |@word trial:10 version:1 compression:1 seems:1 hippocampus:1 seek:1 linearized:2 propagate:1 thereby:1 initial:1 paw:1 efficacy:1 interestingly:1 past:3 current:2 happen:1 informative:4 plasticity:2 motor:3 plot:2 update:1 cue:5 generative:7 nervous:1 realism:1 short:1 sudden:1 provides:1 node:2 ron:10 lx:2 direct:1 differential:1 lx2:1 combine:2 inside:1 pairwise:2 inter:1 indeed:3 behavior:1 isi:4 integrator:2 brain:2 inspired:1 discounted:1 freeman:1 prolonged:1 unpredictable:3 window:1 increasing:1 project:1 estimating:1 moreover:1 underlying:5 notation:1 factorized:1 circuit:1 what:4 emerging:1 certainty:11 temporal:3 act:2 exactly:1 uk:2 unit:1 appear:1 positive:4 before:1 t1:8 local:3 w1k:2 tends:1 limit:1 irreversible:1 consequence:2 despite:1 encoding:3 firing:16 fluctuation:5 black:1 tmax:1 shaded:2 biba:1 range:1 statistically:1 averaged:2 huys:1 acknowledgment:1 practice:1 block:2 implement:6 communicated:1 chaotic:1 signaling:1 area:1 convenient:1 confidence:2 word:3 integrating:1 suggest:1 zoubin:1 get:3 cannot:2 close:3 influence:2 instability:1 accumulating:1 deterministic:3 quick:1 yt:1 compensated:1 go:18 overcounting:1 immediately:2 communicating:2 insight:1 rule:2 importantly:1 otl:5 population:2 stability:1 transmit:1 updated:1 homogeneous:1 hypothesis:1 element:4 crossing:1 expensive:2 net2:1 velocity:1 stripe:13 richards:1 predicts:1 observed:4 role:1 tsodyks:1 connected:3 solla:1 decrease:2 balanced:4 leak:2 dynamic:4 raise:1 tight:1 depend:1 predictive:1 purely:1 easily:1 joint:2 represented:2 train:8 describe:1 london:2 effective:2 zemel:1 outside:1 qon:5 whose:1 encoded:1 fluctuates:1 jean:1 otherwise:2 ability:1 statistic:7 noisy:7 itself:2 online:1 xlt:3 net:1 ucl:1 propose:4 interaction:3 interconnected:1 reset:1 adaptation:2 loop:5 fired:2 x1t:7 lxt:6 parent:1 mazurek:1 bron:1 develop:1 ac:1 propagating:3 recurrent:4 ij:1 received:6 eq:1 strong:1 implemented:3 predicted:3 come:1 convention:1 direction:4 laurens:1 thick:1 opinion:1 implementing:2 redistribution:1 require:3 preliminary:1 proposition:1 biological:2 insert:1 exploring:1 around:1 considered:1 predict:2 integrates:1 currently:1 individually:1 correctness:1 create:1 reflects:1 weighted:1 rough:1 mit:2 gaussian:1 rather:3 avoid:2 command:1 derived:2 likelihood:1 contrast:2 helpful:1 inference:14 dayan:2 el:3 dependent:1 inaccurate:1 integrated:3 entire:4 hidden:14 spurious:1 reproduce:1 france:1 quasi:1 semantics:2 issue:2 overall:3 among:1 constrained:1 integration:7 lif:2 marginal:2 field:1 once:2 having:2 identical:2 represents:2 unsupervised:1 thin:1 report:3 stimulus:4 t2:8 micro:1 few:1 retina:1 consisting:2 fire:6 organization:1 message:1 wc1n:1 implication:1 chain:6 closer:1 necessary:1 incomplete:1 loosely:1 old:2 walk:1 circle:1 plotted:5 causal:1 column:1 rao:2 ar:1 newsome:2 loopy:1 expects:1 rare:1 too:1 optimally:1 reported:1 combined:1 thanks:2 st:5 stay:1 probabilistic:5 off:3 decoding:5 together:1 again:2 central:1 reflect:1 external:1 cognitive:1 leading:2 account:4 potential:7 coding:5 infinitum:1 depends:5 ad:1 performed:4 view:1 observing:2 roff:10 contribution:2 efficiently:1 t3:5 conveying:1 landscape:1 correspond:1 weak:1 bayesian:24 rectified:3 j6:1 submitted:1 synapsis:1 reach:2 synaptic:34 definition:1 failure:1 raster:1 energy:1 sensorimotor:1 frequency:2 infers:1 back:3 appears:1 exceeded:1 higher:2 dt:26 follow:1 wei:2 synapse:5 leen:1 stage:1 hand:1 receives:1 propagation:5 dependant:1 impulse:1 xkt:8 olshausen:1 building:2 effect:1 roitman:1 brown:2 discounting:1 memoryless:1 iteratively:2 illustrated:1 eg:6 conditionally:2 round:1 indistinguishable:1 during:4 self:5 ambiguous:1 whereby:1 excitation:4 linearizing:1 neocortical:1 latham:1 tn:2 motion:11 marry:1 bring:1 meaning:1 common:1 tending:1 stimulation:1 spiking:11 refractory:1 volume:1 cerebral:1 analog:2 interpretation:2 linking:1 refer:1 cambridge:4 fk:3 fano:1 stable:6 cortex:3 inhibition:7 gt:10 posterior:11 own:2 driven:2 certain:2 binary:3 vt:3 muller:1 seen:1 minimum:1 additional:1 fortunately:1 preceding:1 exquisite:1 converge:1 maximize:1 period:4 redundant:3 signal:3 dashed:1 multiple:1 exceeds:3 long:2 post:1 schematic:1 prediction:7 variant:1 qi:1 maintenance:1 xjt:1 involving:1 vision:1 poisson:9 represent:4 normalization:1 invert:1 irregular:1 receive:6 addition:1 whereas:1 fellowship:1 interval:1 dorothy:1 source:1 pyramidal:1 crucial:1 appropriately:1 ot:9 comment:1 supposing:1 contrary:1 odds:18 emitted:1 presence:6 feedforward:4 switch:4 independence:1 affect:3 architecture:2 perfectly:1 topology:1 reduce:1 idea:3 absent:1 inactive:1 fleet:1 ditterich:1 akin:1 movshon:1 peter:2 proceed:1 cause:1 negligible:1 clear:2 dark:1 discount:1 generate:1 wlk:5 inhibitory:4 alters:1 dotted:2 neuroscience:2 jacques:1 per:1 track:1 independency:2 droulez:1 threshold:5 changing:1 rewriting:1 kept:1 deneve:1 downstream:3 sum:1 hodgkin:1 place:2 almost:3 reporting:1 decision:3 dash:1 miserably:1 activity:2 scene:1 explode:1 argument:1 xtk:1 relatively:1 according:1 membrane:7 describes:1 postsynaptic:1 wi:4 evolves:1 making:2 happens:1 intuitively:1 restricted:2 taken:2 equation:3 previously:5 discus:1 count:1 eventually:2 x2t:3 turn:1 appropriate:2 occurrence:1 existence:1 top:1 graphical:2 testable:1 ghahramani:1 disappear:1 society:1 skt:1 psychophysical:1 question:1 quantity:2 spike:56 receptive:1 said:2 thank:1 simulated:1 whom:1 argue:1 presynaptic:1 code:3 providing:1 ratio:23 balance:3 unfortunately:2 debate:1 negative:1 implementation:1 boltzmann:1 perform:1 neuron:69 observation:2 markov:7 hinton:2 saturates:1 neurobiology:1 inh:1 wkl:4 arbitrary:1 drift:1 inferred:1 metabolically:1 kl:3 required:1 connection:4 s20:1 learned:2 nip:1 ytt:1 able:3 usually:1 substract:1 perception:2 regime:2 royal:1 explanation:1 belief:5 event:5 natural:1 representing:1 sk0:2 temporally:1 picture:1 started:1 britten:1 coupled:4 naive:1 text:1 prior:1 marginalizing:1 relative:1 generation:1 interesting:1 analogy:1 versus:1 integrate:3 consistent:3 s0:3 shadlen:3 editor:2 course:2 x3t:1 excitatory:2 surprisingly:1 last:1 free:1 supported:1 bias:4 institute:1 taking:2 markram:1 slt:4 sparse:1 leaky:4 feedback:7 plain:1 cortical:6 world:3 transition:5 rich:1 computes:1 sensory:8 perpetually:1 jump:3 adaptive:1 far:3 emitting:2 kording:1 approximate:1 preferred:3 status:1 unreliable:1 active:3 assumed:2 alternatively:1 learn:1 nature:2 complex:1 meanwhile:1 european:1 linearly:2 noise:2 knill:1 child:8 convey:1 body:2 neuronal:1 fig:1 gatsby:2 fails:1 explicit:1 exponential:1 xl:2 perceptual:2 communicates:1 xt:27 specific:1 jt:1 showing:2 reverberating:1 recurrently:1 decay:1 physiological:3 evidence:16 sit:3 consist:1 albeit:1 conditioned:1 wolpert:1 lt:14 simply:4 appearance:1 likely:2 visual:3 prevents:1 lewicki:1 monotonic:1 corresponds:1 ma:3 conditional:6 presentation:3 towards:2 eventual:1 change:1 tiger:13 determined:1 typical:1 except:2 sophie:1 averaging:1 infinite:1 college:1 internal:1 arises:1 tested:1 phenomenon:1 ex:1 |
1,890 | 2,717 | Binet-Cauchy Kernels
S.V.N. Vishwanathan, Alexander J. Smola
National ICT Australia, Machine Learning Program, Canberra, ACT 0200, Australia
{SVN.Vishwanathan, Alex.Smola}@nicta.com.au
Abstract
We propose a family of kernels based on the Binet-Cauchy theorem and its extension to Fredholm operators. This includes as special cases all currently known
kernels derived from the behavioral framework, diffusion processes, marginalized
kernels, kernels on graphs, and the kernels on sets arising from the subspace angle
approach. Many of these kernels can be seen as the extrema of a new continuum
of kernel functions, which leads to numerous new special cases. As an application,
we apply the new class of kernels to the problem of clustering of video sequences
with encouraging results.
1
Introduction
Recent years have see a combinatorial explosion of results on kernels for structured and
semi-structured data, including trees, strings, graphs, transducers and dynamical systems
[6, 8, 15, 13]. The fact that these kernels are very specific to the type of discrete data under
consideration is a major cause of confusion to the practitioner. What is required is a) an
unified view of the field and b) a recipe to design new kernels easily.
The present paper takes a step in this direction by unifying these diverse kernels by means
of the Binet-Cauchy theorem. Our point of departure is the work of Wolf and Shashua [17],
or more specifically, their proof that det A? B is a kernel on matrices A, B ? Rm?n . We
extend the results of [17] in the following three ways:
1. There exists an operator-valued equivalent of the Binet-Cauchy theorem.
2. Wolf and Shashua only exploit the Binet-Cauchy theorem for one particular choice of
parameters. It turns out that the continuum of these values corresponds to a large class
of kernels some of which are well known and others which are novel.
3. The Binet-Cauchy theorem can be extended to semirings. This points to a close connection with rational kernels [3].
Outline of the paper: Section 2 contains the main result of the present paper: the definition of Binet-Cauchy kernels and their efficient computation. Subsequently, section 3
discusses a number of special cases, which allows us to recover well known kernel functions. Section 4 applies our derivations to the analysis of video sequences, and we conclude
with a discussion of our results.
2
Binet-Cauchy Kernels
In this section we deal with linear mappings from X = Rn to Y = Rm (typically denoted
by matrices), their coordinate free extensions to Fredholm operators (here Rn and Rm are
replaced by measurable sets), and their extensions to semirings (here addition and multiplication are replaced by an abstract class of symbols (?, ?) with the same distributive
properties).
2.1
The General Composition Formula
We begin by defining compound matrices. They arise by picking subsets of entries of a
matrix and computing their determinants.
Definition 1 (Compound Matrix) Let A ? Rm?n , let q ? min(m, n) and let Iqn = {i =
(i1 , i2 , . . . , iq ) : 1 ? i1 < . . . < iq ? n, ii ? N} and likewise Iqm . Then the compound
matrix of order q is defined as
[Cq (A)]i,j := det(A(ik , jl ))qk,l=1 where i ? Iqn and j ? Iqm .
(1)
Here i, j are assumed to be arranged in lexicographical order.
Theorem 2 (Binet-Cauchy) Let A ? Rl?m and, B ? Rl?n . For q ? min(m, n, l) we
have Cq (A? B) = Cq (A)? Cq (B).
When q = m = n = l we have Cq (A) = det(A) and the Binet-Cauchy theorem becomes
the well known identity det(A? B) = det(A) det(B). On the other hand when q = 1 we
have C1 (A) = A, so Theorem 2 reduces to a tautology.
Theorem 3 (Binet-Cauchy for Semirings) When the common semiring (R, +, ?, 0, 1) is
replaced by an abstract semiring (K, ?, ?, ?0, ?1) the equality Cq (A? B) = Cq (A)? Cq (B)
still holds. Here all operations occur on the monoid K, addition and multiplication are
replaced by ?, ?, and (?
0, ?
1) take the role of (0, 1).
A second extension of Theorem 2 is to replace matrices by Fredholm operators, as they
can be expressed as integral operators with corresponding kernels. In this case, Theorem 2
becomes a statement about convolutions of integral kernels.
Definition 4 (Fredholm Operator) A Fredholm operator is a bounded linear operator between two Hilbert spaces with closed range and whose kernel and co-kernel are finitedimensional.
Theorem 5 (Kernel Representation of Fredholm Operators) Let A : L2 (Y) ? L2 (X)
and, B : L2 (Y) ? L2 (Z) be Fredholm operators. Then there exists some kA : X ? Y ? R
such that for all f ? L2 (X) we have
Z
[Af ](x) =
kA (x, y)f (y)dy.
(2)
Y
?
Moreover, for the composition A B we have kA? B (x, z) =
R
Y
kA? (x, y)kB (y, z)dy.
Here the convolution of kernels kA and kB plays the same role as the matrix multiplication. To extend the Binet-Cauchy theorem we need to introduce the analog of compound
matrices:
Definition 6 (Compound Kernel and Operator) Denote by X, Y ordered sets and let k :
X ? Y ? R. Define IqX = {x ? Xq : x1 ? . . . ? xq } and likewise IqY . Then the compound
kernel of order q is defined as
k [q] (x, y) := det(k(xk , yl ))qk,l=1 where x ? IqX and y ? IqY .
(3)
If k is the integral kernel of an operator A we define Cq (A) to be the integral operator
corresponding to k [q] .
Theorem 7 (General Composition Formula [11]) Let X, Y, Z be ordered sets and let A :
L2 (Y) ? L2 (X), B : L2 (Y) ? L2 (Z) be Fredholm operators. Then for q ? N we have
Cq (A? B) = Cq (A)? Cq (B).
To recover Theorem 2 from Theorem 7 set X = [1..m], Y = [1..n] and Z = [1..l].
(4)
2.2
Kernels
The key idea in turning the Binet-Cauchy theorem and its various incarnations into a kernel
is to exploit the fact that tr A? B and det A? B are kernels on operators A, B. We extend
this by replacing A? B with some functions ?(A)? ?(B) involving compound operators.
Theorem 8 (Trace and Determinant Kernel) Let A, B : L2 (X) ? L2 (Y) be Fredholm
operators and let S : L2 (Y) ? L2 (Y), T : L2 (X) ? L2 (X) be positive trace-class operators. Then the following two kernels are well defined and they satisfy Mercer?s condition:
k(A, B) = tr SA? T B
(5)
?
k(A, B) = det SA T B .
(6)
Note that determinants are not defined in general for infinite dimensional operators, hence
our restriction to Fredholm operators A, B in (6).
Proof Observe that S and T are positive and compact. Hence they admit a decomposition
?
into S = VS VS? and
virtue of the commutativity of the trace we have
T = VT VT . By
?
that k(A, B) = tr [VT AVS ] [VT BVS ] . Analogously, using the Binet-Cauchy theorem,
we can decompose the determinant. The remaining terms VT AVS and VT BVS are again
Fredholm operators for which determinants are well defined.
Next we use special choices of A, B, S, T involving compound operators directly to state
the main theorem of our paper.
Theorem 9 (Binet-Cauchy Kernel) Under the assumptions
of Theorem 8 it follows
that
for all q ? N the kernels k(A, B) = tr Cq SA? T B and k(A, B) = det Cq SA? T B
satisfy Mercer?s condition.
Proof We exploit the factorization S = VS VS? , T = VT? VT and apply Theorem 7. This
yields Cq (SA? T B) = Cq (VT AVS )? Cq (VT BVS ), which proves the theorem.
Finally, we define a kernel based on the Fredholm determinant itself. It is essentially a
weighted combination of Binet-Cauchy kernels. Fredholm determinants are defined as
follows [11]:
?
X
?q
tr Cq (A).
(7)
D(A, ?) :=
q!
q=1
This series converges for all ? ? C and it is an entire function of ?. It suggests a kernel
involving weighted combinations of the kernels of Theorem 9. We have the following:
Corollary 10 (Fredholm Kernel) Let A, B, S, T as in Theorem 9 and let ? > 0. Then the
following kernel satisfies Mercer?s condition:
k(A, B) := D(A? B, ?) where ? > 0.
(8)
?
D(A B, ?) is a weighted combination of the kernels discussed above. The exponential
1
down-weighting via q!
ensures that the series converges even in the case of exponential
growth of the values of the compound kernel.
2.3
Efficient Computation
At first glance, computing the kernels of Theorem 9 and Corollary 10 presents a formidable
m?n
computational task
, the matrix
even in the finite dimensional case. If A, B ? R
n
?
Cq (A B) has q rows and columns and each of the entries requires the computation of
a determinant of a q-dimensional matrix. A brute-force approach would involve O(q 3 nq )
operations (assuming 2q ? n). Clearly we need more efficient techniques.
When computing determinants, we can take recourse to Franke?s Theorem [7] which states
that
n?1
det Cq (A) = (det A)( q?1 ) .
(9)
n?1
and consequently k(A, B) = det Cq [SA? T B] = (det[SA? T B])( q?1 ) .1 This indicates
that the determinant kernel may be of limited use, due to the typically quite high power in
the exponent. Kernels building on tr Cq are not plagued by this problem and we give an
efficient recursion below. It follows from the ANOVA kernel recursion of [1]:
Lemma 11 Denote by A ? Cm?m a square matrix and let ?1 , . . . , ?m be its eigenvalues.
Then tr Cq (A) can be computed by the following recursion:
q
tr Cq (A) =
n
X q
1X
(?1)j+1 C?q?j (A)C?j (A) where C?q (A) =
?j .
q j=1
j=1
(10)
Proof We begin by writing A in its Jordan normal form as A = P DP ?1 where D is a
block diagonal, upper triangular matrix. Furthermore, the diagonal elements of D consist
of the eigenvalues of A. Repeated application of the Binet-Cauchy Theorem yields
tr Cq (A) = tr Cq (P )Cq (D)Cq (P ?1 ) = tr Cq (D)Cq (P ?1 )Cq (P ) = tr Cq (D)
(11)
For a triangular matrix the determinant is the product of its diagonal entries. Since all
the square submatrices of D are also upper triangular, to construct tr(Cq (D)) we need to
sum over all products of exactly q eigenvalues. This is analog to the requirement of the
ANOVA kernel of [1]. In its simplified version it can be written as (10), which completes
the proof.
We can now compute the Jordan normal form of SA? T B in O(n3 ) time and apply
Lemma 11 directly to it to compute the kernel value.
Finally, in the case of Fredholm determinants, we can use the recursion directly, because
for n-dimensional matrices the sum terminates after n terms. This is no more expensive
than computing tr Cq directly. Note that in the general nonsymmetric case (i.e. A 6= A? )
no such efficient recursions are known.
3
Special Cases
We now focus our attention on various special cases to show how they fit into the general
framework which we developed in the previous section. For this to succeed, we will map
various systems such as graphs, dynamical systems, or video sequences into Fredholm
operators. A suitable choice of this mapping and of the operators S, T of Theorem 9 will
allow us to recover many well-known kernels as special cases.
3.1
Dynamical Systems
We begin by describing a partially observable discrete time LTI (Linear Time Invariant)
model commonly used in control theory. Its time-evolution equations are given by
yt = P xt + wt
xt = Qxt?1 + vt
where wt ? N(0, R)
where vt ? N(0, S).
(12a)
(12b)
Here yt ? Rm is observed, xt ? Rn is the hidden or latent variable, and P ? Rm?n , Q ?
Rn?n , R ? Rm?m and, S ? Rn?n , moreover R, S 0. Typically m ? n. similar model
exists for continuous LTI. Further details on it can be found in [14].
Following the behavioral framework of [16] we associate dynamical systems, X :=
(P, Q, R, S, x0 ), with their trajectories, that is, the set of yt with t ? N for discrete time
systems (and t ? [0, ?) for the continuous-time case). These trajectories can be interpreted
1
Eq. (9) can be seen as follows: the compound matrix of an orthogonal matrix is orthogonal and
consequently its determinant is unity. Subsequently use an SVD factorization of the argument of the
compound operator to compute the determinant of the compound matrix of a diagonal matrix.
as linear operators mapping from Rm (the space of observations y) into the time domain
(N or [0, ?)) and vice versa. The diagram below depicts this mapping:
X
/ Traj(X)
/ Cq (Traj(X))
Finally, Cq (Traj(X)) is weighted in a suitable fashion by operators S and T and the trace
is evaluated. This yields an element from the family of Binet-Cauchy kernels.
In the following we discuss several kernels and we show that they differ essentially
in how the mapping into a dynamical system occurs (discrete-time or continuous time,
fully observed or partial observations), whether any other preprocessing is carried out on
Cq (Traj(X)) (such as QR decomposition in the case of the kernel proposed by [10] and
rediscovered by [17]), or which weighting S, T is chosen.
3.2
Dynamical Systems Kernels
We begin with kernels on dynamical systems (12) as proposed in [14]. Set S = 1, q = 1
and T to be the diagonal operator with entries e??t . In this case the Binet-Cauchy kernel
between systems X and X ? becomes
?
X
tr Cq (S Traj(X) T Traj(X ? )? ) =
e??t yt? yt? .
(13)
i=1
Since yt , yt? are random variables, we also need to take expectations over wt , vt , wt? , vt? .
Some tedious yet straightforward algebra [14] allows us to compute (13) as follows:
1
?
tr [SM2 + R] ,
(14)
k(X, X ? ) = x?
0 M1 x0 + ?
e ?1
where M1 , M2 satisfy the Sylvester equations:
M1 = e?? Q? P ? P ? Q? + e?? Q? M1 Q? and M2 = P ? P ? + e?? Q? M2 Q? .
(15)
3
Such kernels can be computed in O(n ) time [5]. Analogous expressions for continuoustime systems exist [14]. In Section 4 we will use this kernel to compute similarities between
video sequences, after having encoded the latter as a dynamical system. This will allow us
to compare sequences of different length, as they are all mapped to dynamical systems in
the first place.
3.3
Martin Kernel
A characteristic property of (14) is that it takes initial conditions of the dynamical system
into account. If this is not desired, one may choose to pick only the subspace spanned by
the trajectory yt . This is what was proposed in [10].2
More specifically, set S = T = 1, consider the trajectory upto only a finite number of time
steps, say up to n, and let q = n. Furthermore let Traj(X) = QX RX denote the QRdecomposition of Traj(X), where QX is an orthogonal matrix and RX is upper triangular.
Then it can be easily verified the kernel proposed by [10] can be written as
?
k(X, X ? ) = tr Cq (SQX T Q?
X ? ) = det(QX QX ? ).
(16)
This similarity measure has been used by Soatto, Doretto, and coworkers [4] for the analysis and computation of similarity in video sequences. Subsequently Wolf and Shashua
[17] modified (16) to allow for kernels: to deal with determinants on a possibly infinitedimensional feature space they simply project the trajectories on a reduced set of points in
feature space.3 This is what [17] refer to as a kernel on sets.
2
Martin [10] suggested the use of Cepstrum coefficients of a dynamical system to define a Euclidean metric. Later De Cock and Moor [2] showed that this distance is, indeed, given by the
computation of subspace angles, which can be achieved by computing the QR-decomposition.
3
To be precise, [17] are unaware of the work of [10] or of [2] and they rediscover the notion of
subspace angles for the purpose of similarity measures.
3.4
Graph Kernels
Yet another class of kernels can be seen to fall into this category: the graph kernels proposed
in [6, 13, 9, 8]. Denote by G(V, E) a graph with vertices V and edges E. In some cases,
such as in the analysis of molecules, the vertices will be equipped with labels L. For
recovering these kernels from our framework we set q = 1 and systematically map graph
kernels to dynamical systems.
We denote by xt a probability distribution over the set of vertices at time t. The timeevolution xt ? xt+1 occurs by performing a random walk on the graph G(V, E). This
yields xt+1 = W D?1 xt , where W is the connectivity matrix of the graph and D is a
diagonal matrix where Dii denotes the outdegree of vertex i. For continuous-time systems
?
? is the normalized graph Laplacian [9].
one uses x(t) = exp(?Lt)x(0),
where L
In the graph kernels of [9, 13] one assumes that the variables xt are directly observed and
no special mapping is required in order to obtain yt . Various choices of S and T yield the
following kernels:
? [9] consider a snapshot of the diffusion process at t = ? . This amounts to choosing
T = 1 and a S which is zero except for a diagonal entry at ? .
? The inverse Graph-Laplacian kernel proposed in [13] uses a weighted combination of
diffusion process and corresponds to S = W a diagonal weight matrix.
? The model proposed in [6] can be seen as using a partially observable model: rather
than observing the states directly, we only observe the labels emitted at the states. If we
associate this mapping from states to labels with the matrix P of (12), set T = 1 and
let S be the projector on the first n time instances, we recover the kernels from [6].
So far, we deliberately made no distinction between kernels on graphs and kernels between
graphs. This is for good reason: the trajectories depend on both initial conditions and the
dynamical system itself. Consequently, whenever we want to consider kernels between initial conditions, we choose the same dynamical system in both cases. Conversely, whenever
we want to consider kernels between dynamical systems, we average over initial conditions.
This is what allows us to cover all the aforementioned kernels in one framework.
3.5
Extensions
Obviously the aforementioned kernels are just specific instances of what is possible by
using kernels of Theorem 9. While it is pretty much impossible to enumerate all combinations, we give a list of suggestions for possible kernels below:
? Use the continuous-time diffusion process and a partially observable model. This would
extend the diffusion kernels of [9] to comparisons between vertices of a labeled graph
(e.g. atoms in a molecule).
? Use diffusion processes to define similarity measures between graphs.
? Compute the determinant of the trajectory associated with an n-step random walk on a
graph, that is use Cq with q = n instead of C1 . This gives a kernel analogous to the
one proposed by Wolf and Shashua [17], however without the whitening incurred by
the QR factorization.
? Take Fredholm determinants of the above mentioned trajectories.
? Use a nonlinear version of the dynamical system as described in [14].
4
Experiments
To test the utility of our kernels we applied it to the task of clustering short video clips.
We randomly sampled 480 short clips from the movie Kill Bill and model them as linear
ARMA models (see Section 3.1).
The sub-optimal procedure outlined in [4] was used
for estimating the model parameters P , Q, R and,
S and the kernels described in Section 3.2 were applied to these models. Locally Linear Embedding
(LLE) [12] was used to cluster and embed the clips
in two dimensions. The two dimensional embedding obtained by LLE is depicted in Figure 1. We
randomly selected a few data points from Figure 1
and depict the first frame of the corresponding clips
in Figure 2.
Observe the linear cluster (with a projecting arm) in
Figure 1: LLE embeddings of 480 ran- Figure 1. This corresponds to clips which are temdom clips from Kill Bill
porally close to each other and hence have similar
dynamics.
For
instance
clips
in
the
far
right
depict
a
person rolling in the snow while those in the far left corner depict a sword fight while clips
in the center involve conversations between two characters. A naiv?e comparison of the intensity values or a dot product of the actual clips would not be able to extract such semantic
information. Even though the camera angle varies with time our kernel is able to successfully pick out the underlying dynamics of the scene. These experiments are encouraging
and future work will concentrate on applying this to video sequence querying.
2
1
0
?1
?2
?3
?4
?5
?4
?3
?2
?1
0
1
2
3
4
Figure 2: LLE embeddings of a subset of our dataset. A larger version is available from
http://mlg.anu.edu.au/?vishy/papers/KillBill.png
5
Discussion
In this paper, we introduced a unifying framework for defining kernels on discrete objects
using the Binet-Cauchy theorem on compounds of the Fredholm operators. We demonstrated that many of the previously known kernels can be explained neatly by our framework. In particular many graph kernels and dynamical system related kernels fall out as
natural special cases. The main advantage of our unifying framework is that it allows kernel engineers to use domain knowledge in a principled way to design kernels for solving
real life problems.
Acknowledgement We thank Stephane Canu and Ren?e Vidal for useful discussions. National ICT Australia is supported by the Australian Government?s Program Backing Australia?s Ability. This work was partly supported by grants of the Australian Research Council. This work was supported by the IST Programme of the European Community, under
the Pascal Network of Excellence, IST-2002-506778.
References
[1] C. J. C. Burges and V. Vapnik. A new method for constructing artificial neural networks. Interim technical report, ONR contract N00014 - 94-c-0186, AT&T Bell Laboratories, 1995.
[2] K. De Cock and B. De Moor. Subspace angles between ARMA models. Systems and
Control Letter, 46:265 ? 270, 2002.
[3] C. Cortes, P. Haffner, and M. Mohri. Rational kernels. In Proceedings of Neural
Information Processing Systems 2002, 2002. in press.
[4] G. Doretto, A. Chiuso, Y.N. Wu, and S. Soatto. Dynamic textures. International
Journal of Computer Vision, 51(2):91 ? 109, 2003.
[5] J. D. Gardiner, A. L. Laub, J. J. Amato, and C. B. Moler. Solution of the Sylvester matrix equation AXB ? + CXD? = E. ACM Transactions on Mathematical Software,
18(2):223 ? 231, 1992.
[6] T. G?artner, P.A. Flach, and S. Wrobel. On graph kernels: Hardness results and efficient alternatives. In B. Sch?olkopf and M. Warmuth, editors, Sixteenth Annual Conference on Computational Learning Theory and Seventh Kernel Workshop, COLT.
Springer, 2003.
[7] W. Gr?obner. Matrizenrechnung. BI Hochschultaschenb?ucher, 1965.
[8] H. Kashima, K. Tsuda, and A. Inokuchi. Marginalized kernels between labeled
graphs. In Proceedings of the 20th International Conference on Machine Learning
(ICML), Washington, DC, United States, 2003.
[9] I.R. Kondor and J. D. Lafferty. Diffusion kernels on graphs and other discrete structures. In Proceedings of the ICML, 2002.
[10] R.J. Martin. A metric for ARMA processes. IEEE Transactions on Signal Processing,
48(4):1164 ? 1170, 2000.
[11] A. Pinkus. Spectral properties of totally positive kernels and matrices. In M. Gasca
and C. A. Miccheli, editors, Total Positivity and its Applications, volume 359 of Mathematics and its Applications, pages 1?35. Kluwer, March 1996.
[12] S. Roweis and L. K. Saul. Nonlinear dimensionality reduction by locally linear embedding. Science, 290:2323 ? 2326, December 2000.
[13] A.J. Smola and I.R. Kondor. Kernels and regularization on graphs. In B. Sch?olkopf
and M. K. Warmuth, editors, Proceedings of the Annual Conference on Computational Learning Theory, Lecture Notes in Computer Science. Springer, 2003.
[14] A.J. Smola, R. Vidal, and S.V.N. Vishwanathan. Kernels and dynamical systems.
Automatica, 2004. submitted.
[15] S.V.N. Vishwanathan and A.J. Smola. Fast kernels on strings and trees. In Proceedings of Neural Information Processing Systems 2002, 2002.
[16] J. C. Willems. From time series to linear system. I. Finite-dimensional linear time
invariant systems. Automatica J. IFAC, 22(5):561 ? 580, 1986.
[17] L. Wolf and A. Shashua. Learning over sets using kernel principal angles. Jounal of
Machine Learning Research, 4:913 ? 931, 2003.
| 2717 |@word determinant:17 kondor:2 version:3 flach:1 tedious:1 decomposition:3 pick:2 tr:17 incarnation:1 reduction:1 initial:4 contains:1 series:3 united:1 ka:5 com:1 yet:2 written:2 depict:3 v:4 selected:1 nq:1 warmuth:2 xk:1 short:2 mathematical:1 ik:1 chiuso:1 transducer:1 laub:1 artner:1 behavioral:2 introduce:1 excellence:1 x0:2 hardness:1 indeed:1 encouraging:2 actual:1 equipped:1 totally:1 becomes:3 begin:4 project:1 bounded:1 moreover:2 formidable:1 estimating:1 underlying:1 what:5 cm:1 interpreted:1 string:2 developed:1 unified:1 extremum:1 monoid:1 act:1 growth:1 exactly:1 rm:8 brute:1 control:2 grant:1 positive:3 au:2 suggests:1 conversely:1 co:1 factorization:3 limited:1 range:1 bi:1 lexicographical:1 camera:1 block:1 procedure:1 submatrices:1 bell:1 close:2 operator:29 impossible:1 applying:1 writing:1 franke:1 restriction:1 equivalent:1 measurable:1 map:2 yt:9 projector:1 bill:2 straightforward:1 attention:1 center:1 demonstrated:1 m2:3 spanned:1 embedding:3 notion:1 coordinate:1 analogous:2 play:1 us:2 associate:2 element:2 expensive:1 labeled:2 observed:3 role:2 ensures:1 ran:1 mentioned:1 principled:1 dynamic:3 depend:1 solving:1 algebra:1 easily:2 various:4 derivation:1 fast:1 artificial:1 choosing:1 whose:1 quite:1 encoded:1 valued:1 larger:1 say:1 triangular:4 ability:1 itself:2 obviously:1 sequence:7 eigenvalue:3 advantage:1 propose:1 product:3 roweis:1 sixteenth:1 olkopf:2 qr:3 recipe:1 cluster:2 requirement:1 converges:2 object:1 iq:2 eq:1 sa:8 recovering:1 australian:2 differ:1 direction:1 snow:1 concentrate:1 stephane:1 subsequently:3 kb:2 australia:4 dii:1 government:1 decompose:1 extension:5 hold:1 bvs:3 plagued:1 normal:2 exp:1 mapping:7 major:1 continuum:2 purpose:1 combinatorial:1 currently:1 label:3 council:1 vice:1 successfully:1 weighted:5 moor:2 clearly:1 modified:1 rather:1 corollary:2 derived:1 focus:1 amato:1 indicates:1 cock:2 typically:3 entire:1 fight:1 hidden:1 i1:2 backing:1 aforementioned:2 colt:1 pascal:1 denoted:1 exponent:1 special:9 field:1 construct:1 having:1 washington:1 atom:1 outdegree:1 icml:2 future:1 others:1 report:1 few:1 randomly:2 national:2 replaced:4 continuoustime:1 rediscovered:1 semirings:3 integral:4 edge:1 explosion:1 partial:1 orthogonal:3 tree:2 commutativity:1 euclidean:1 walk:2 desired:1 arma:3 tsuda:1 instance:3 column:1 cover:1 vertex:5 subset:2 entry:5 rolling:1 seventh:1 gr:1 varies:1 iqn:2 person:1 international:2 contract:1 yl:1 picking:1 analogously:1 connectivity:1 again:1 choose:2 sqx:1 possibly:1 positivity:1 admit:1 corner:1 cxd:1 account:1 de:3 includes:1 coefficient:1 satisfy:3 later:1 view:1 closed:1 observing:1 shashua:5 recover:4 square:2 qk:2 characteristic:1 likewise:2 yield:5 fredholm:18 ren:1 trajectory:8 rx:2 submitted:1 whenever:2 definition:4 mlg:1 vishy:1 iqx:2 proof:5 associated:1 rational:2 sampled:1 dataset:1 conversation:1 knowledge:1 dimensionality:1 hilbert:1 cepstrum:1 arranged:1 evaluated:1 though:1 furthermore:2 just:1 smola:5 hand:1 replacing:1 nonlinear:2 glance:1 building:1 normalized:1 binet:20 deliberately:1 evolution:1 regularization:1 hence:3 equality:1 soatto:2 laboratory:1 i2:1 semantic:1 deal:2 outline:1 confusion:1 consideration:1 novel:1 common:1 rl:2 pinkus:1 jl:1 extend:4 analog:2 discussed:1 m1:4 nonsymmetric:1 volume:1 kluwer:1 refer:1 composition:3 versa:1 outlined:1 canu:1 mathematics:1 neatly:1 dot:1 similarity:5 whitening:1 recent:1 showed:1 compound:13 n00014:1 onr:1 moler:1 vt:14 life:1 seen:4 doretto:2 coworkers:1 signal:1 semi:1 ii:1 reduces:1 technical:1 ifac:1 af:1 laplacian:2 involving:3 sylvester:2 essentially:2 expectation:1 metric:2 vision:1 kernel:100 achieved:1 c1:2 addition:2 want:2 completes:1 diagram:1 sch:2 december:1 lafferty:1 jordan:2 practitioner:1 emitted:1 embeddings:2 fit:1 iqm:2 idea:1 haffner:1 svn:1 det:15 whether:1 expression:1 utility:1 cause:1 enumerate:1 useful:1 involve:2 amount:1 locally:2 clip:9 png:1 category:1 reduced:1 http:1 exist:1 arising:1 kill:2 diverse:1 discrete:6 ist:2 key:1 anova:2 verified:1 lti:2 diffusion:7 graph:22 year:1 sum:2 angle:6 inverse:1 letter:1 place:1 family:2 wu:1 dy:2 annual:2 gardiner:1 occur:1 vishwanathan:4 alex:1 n3:1 scene:1 software:1 argument:1 min:2 performing:1 interim:1 martin:3 structured:2 combination:5 march:1 terminates:1 character:1 unity:1 projecting:1 invariant:2 explained:1 recourse:1 equation:3 previously:1 turn:1 discus:2 describing:1 tautology:1 available:1 operation:2 vidal:2 apply:3 observe:3 upto:1 spectral:1 kashima:1 alternative:1 denotes:1 remaining:1 clustering:2 assumes:1 marginalized:2 unifying:3 sm2:1 exploit:3 prof:1 occurs:2 diagonal:8 dp:1 subspace:5 distance:1 thank:1 mapped:1 distributive:1 cauchy:20 reason:1 nicta:1 assuming:1 length:1 cq:40 statement:1 trace:4 design:2 upper:3 av:3 convolution:2 observation:2 snapshot:1 willems:1 finite:3 defining:2 extended:1 precise:1 frame:1 rn:5 dc:1 community:1 intensity:1 introduced:1 semiring:2 required:2 connection:1 distinction:1 able:2 suggested:1 dynamical:18 below:3 departure:1 iqy:2 program:2 including:1 video:7 power:1 suitable:2 natural:1 force:1 turning:1 qxt:1 recursion:5 arm:1 movie:1 numerous:1 carried:1 extract:1 xq:2 ict:2 l2:15 acknowledgement:1 multiplication:3 fully:1 lecture:1 suggestion:1 querying:1 incurred:1 sword:1 mercer:3 editor:3 systematically:1 row:1 mohri:1 supported:3 free:1 allow:3 lle:4 burges:1 fall:2 saul:1 dimension:1 finitedimensional:1 unaware:1 infinitedimensional:1 commonly:1 made:1 preprocessing:1 simplified:1 programme:1 far:3 qx:4 transaction:2 compact:1 observable:3 automatica:2 conclude:1 assumed:1 continuous:5 latent:1 pretty:1 molecule:2 traj:8 european:1 constructing:1 domain:2 main:3 arise:1 inokuchi:1 repeated:1 x1:1 canberra:1 depicts:1 fashion:1 sub:1 exponential:2 weighting:2 theorem:32 formula:2 down:1 embed:1 specific:2 xt:9 wrobel:1 symbol:1 list:1 cortes:1 virtue:1 exists:3 consist:1 workshop:1 vapnik:1 texture:1 anu:1 depicted:1 lt:1 simply:1 rediscover:1 expressed:1 ordered:2 partially:3 applies:1 springer:2 wolf:5 corresponds:3 satisfies:1 acm:1 succeed:1 identity:1 consequently:3 replace:1 axb:1 specifically:2 infinite:1 except:1 wt:4 lemma:2 engineer:1 total:1 principal:1 partly:1 svd:1 latter:1 alexander:1 |
1,891 | 2,718 | Semi-supervised Learning on Directed Graphs
Dengyong Zhou? , Bernhard Sch?olkopf? , and Thomas Hofmann??
?
Max Planck Institute for Biological Cybernetics
72076 Tuebingen, Germany
{dengyong.zhou, bernhard.schoelkopf}@tuebingen.mpg.de
?
Department of Computer Science, Brown University
Providence, RI 02912 USA
[email protected]
Abstract
Given a directed graph in which some of the nodes are labeled, we investigate the question of how to exploit the link structure of the graph to infer
the labels of the remaining unlabeled nodes. To that extent we propose a
regularization framework for functions defined over nodes of a directed
graph that forces the classification function to change slowly on densely
linked subgraphs. A powerful, yet computationally simple classification
algorithm is derived within the proposed framework. The experimental
evaluation on real-world Web classification problems demonstrates encouraging results that validate our approach.
1
Introduction
We consider semi-supervised classification problems on weighted directed graphs, in which
some nodes in the graph are labeled as positive or negative, and where the task consists in
classifying unlabeled nodes. Typical examples of this kind are Web page categorization
based on hyperlink structure [4, 11] and document classification or recommendation based
on citation graphs [10], yet similar problems exist in other domains such as computational
biology. For the sake of concreteness, we will mainly focus on the Web graph in the sequel,
i.e. the considered graph represents a subgraph of the Web, where nodes correspond to Web
pages and directed edges represent hyperlinks between them (cf. [3]).
We refrain from utilizing attributes or features associated with each node, which may or
may not be available in applications, but rather focus on the analysis of the connectivity of
the graph as a means for classifying unlabeled nodes. Such an approach inevitably needs
to make some a priori premise about how connectivity and categorization of individual
nodes may be related in real-world graphs. The fundamental assumption of our framework
is the category similarity of co-linked nodes in a directed graph. This is a slightly more
complex concept than in the case of undirected (weighted) graphs [1, 18, 12, 15, 17], where
a typical assumption is that an edge connecting two nodes will more or less increase the
likelihood of the nodes belonging to the same category. Co-linkage on the other hand
seems a more suitable and promising concept in directed graphs, as is witnessed by its
successful use in Web page categorization [4] as well as co-citation analysis for information
retrieval [10]. Notice that co-linkage comes in two flavors: sibling structures, i.e. nodes
with common parents, and co-parent structures, i.e. nodes with common children. In most
Web and citation graph related application, the first assumption, namely that nodes with
highly overlapping parent sets are likely to belong to the same category, seems to be more
relevant (cf. [4]), but in general this will depend on the specific application.
One possible way of designing classifiers based on graph connectivity is to construct a
kernel matrix based on pairwise links [11] and then to adopt a standard kernel method,
e.g. Support Vector Machines (SVMs) [16] as a learning algorithm. However, a kernel
matrix as the one proposed in [11] only represents local relationships among nodes, but
completely ignores the global structure of the graph. The idea of exploiting global rather
than local graph structure is widely used in other Web-related techniques, including Web
page ranking [2, 13], finding similar Web pages [7], detecting Web communities [13, 9]
and so on. The major innovation of this paper is a general regularization framework on
directed graphs, in which the directionality and global relationships are considered, and
a computationally attractive classification algorithm, which is derived from the proposed
regularization framework.
2
2.1
Regularization Framework
Preliminaries
A directed graph ? = (V, E) consists of a set of vertices, denoted by V and a set of edges,
denoted by E ? V ? V . Each edge is an ordered pair of nodes [u, v] representing a
directed connection from u to v. We do not allow self loops, i.e. [v, v] 6? E for all v ? V .
In a weighted directed graph, a weight function w : V ? V ? R+ is associated with ?,
satisfying w([u, v]) = 0 if and only if [u, v] 6? E. Typically, we can equip a directed graph
with a canonical weight function by defining w([u, v]) ? 1 if and only if [u, v] ? E. The
in-degree p(v) and out-degree q(v) of a vertex v ? V , respectively, are defined as
X
X
p(v) ?
w([u, v]), and q(v) ?
w([v, u]) .
(1)
{u|[u,v]?E}
{u|[v,u]?E}
Let H(V ) denote the space of functions f : V ? R, which assigns a real value f (v) to
each vertex v. The function f can be represented as a column vector in R|V | , where |V |
denotes the number of the vertices in V . The function space H(V ) can be endowed with
the usual inner product:
X
hf, gi =
f (v)g(v).
(2)
v
Accordingly, the norm of the function induced from the inner product is kf k =
2.2
p
hf, f i.
Bipartite Graphs
A bipartite graph G = (H, A, L) is a special type of directed graph that consists of two
sets of vertices, denoted by H and A respectively, and a set of edges (or links), denoted by
L ? H ? A. In a bipartite graph, each edge connects a vertex in H to a vertex in A. Any
directed graph ? = (V, E) can be regarded as a bipartite graph using the following simple
construction [14]: H ? {h|h ? V, q(h) > 0}, A ? {a|a ? V, p(a) > 0}, and L ? E.
Figure 1 depicts the construction of the bipartite graph. Notice that vertices of the original
graph ? may appear in both vertex sets H and A of the constructed bipartite graph.
The intuition behind the construction of the bipartite graph is provided by the so-called hub
and authority web model introduced by Kleinberg [13]. The model distinguishes between
two types of Web pages: authoritative pages, which are pages relevant to some topic,
and hub pages, which are pages pointing to relevant pages. Note that some Web pages can
Figure 1: Constructing a bipartite graph from a directed one. Left: directed graph. Right:
bipartite graph. The hub set H = {1, 3, 5, 6}, and the authority set A = {2, 3, 4}. Notice
that the vertex indexed by 3 is simultaneously in the hub and authority set.
simultaneously be both hub and authority pages (see Figure 1). Hubs and authorities exhibit
a mutually reinforcement relationship: a good hub node points to many good authorities
and a good authority node is pointed to by many good hubs. It is interesting to note that in
general there is no direct link from one authority to another. It is the hub pages that glue
together authorities on a common topic.
According to Kleinberg?s model, we suggestively call the vertex set H in the bipartite graph
the hub set, and the vertex set A the authority set.
2.3
Smoothness Functionals
If two distinct vertices u and v in the authority set A are co-linked by vertex h in the hub
set H as shown in the left panel of Figure 2, then we think that u and v are likely to be
related, and the co-linkage strength induced by h between u and v can be measured by
ch ([u, v]) =
w([h, u])w([h, v])
.
q(h)
(3)
In addition, we define ch (v, v) = 0 for all v in the authority set A and for all h in the hub
set H. Such a relevance measure can be naturally understood in the situation of citation
networks. If two articles are simultaneously cited by some other article, then this should
make it more likely that both articles deal with a similar topics. Moreover, the more articles
cite both articles together, the more significant the connection. A natural question arising
in this context is why the relevance measure is further normalized by out-degree. Let us
consider the following two web sites: Yahoo! and kernel machines. General interest portals
like Yahoo! consists of pages having a large number of diverse hyperlinks. The fact that
two web pages are co-linked by Yahoo! does not establish a significant connection between
them. In contrast, the pages on the kernel machine Web site have much fewer hyperlinks,
but the Web pages pointed to are closely related in topic.
Let f denote a function defined on the authority set A. The smoothness of function f can
be measured by the following functional:
2
f (u)
f (v)
1 XX
ch ([u, v]) p
?p
.
(4)
?A (f ) =
2 u,v
p(u)
p(v)
h
The smoothness functional penalizes large differences in function values for vertices in the
authority set A that are strongly related. Notice that the function values are normalized by
Figure 2: Link and relevance. Left panel: vertices u and v in the authority set A are colinked by vertex h in the hub set H. Right panel: vertices u and v in the hub set H co-link
vertex a in the authority set A.
in-degree. For the Web graph, the explanation is similar to the one given before. Many
web pages contain links to popular sites like the Google search engine. This does not mean
though that all these Web pages share a common topic. However, if two web pages point to
web page like the one of the Learning with Kernels book, it is likely to express a common
interest for kernel methods.
Now define a linear operator T : H(A) ? H(H) by
X w([h, a])
p
f (a).
(T f )(h) =
q(h)p(a)
a
Then its adjoint T ? : H(H) ? H(A) is given by
X w([h, a])
p
(T ? f )(a) =
f (h).
q(h)p(a)
h
(5)
(6)
These two operators T and T ? were also implicitly suggested by [8] for developing a new
Web page ranking algorithm. Further define the operator SA : H(A) ? H(A) by composing T and T ? , i.e.
SA = T ? T,
(7)
and the operator ?A : H(A) ? H(A) by
?A = I ? S A ,
(8)
where I denotes the identity operator. Then we can show the following (See Appendix A
for the proof):
Proposition 1. ?A (f ) = hf, ?A f i.
Comparing with the combinatorial Laplace operator defined on undirected graphs [5], we
can think of the operator ?A as a Laplacian but defined on the authority set of directed
graphs. Note that Proposition 1 also shows that the Laplacian ?A is positive semi-definite.
In fact, we can further show that the eigenvalues of the operator SA are scattered in [0, 1],
and accordingly the eigenvalues of the Laplacian ?A fall into [0, 1].
Similarly, if two distinct vertices u and v co-link vertex a in the authority set A as shown in
right panel of Figure 2, then u and v are also thought to be related. The co-linkage strength
between u and v induced by a can be measured by
ca ([u, v]) =
w([u, a])w([v, a])
.
p(a)
(9)
and the smoothness of function f on the hub set H can be measured by:
2
1 XX
f (u)
f (v)
?H (f ) =
ca ([u, v]) p
?p
.
2 u,v a
q(u)
q(v)
(10)
As before, one can define the operators SH = T T ? and ?H = I ? SH leading to the
corresponding statement:
Proposition 2. ?H (f ) = hf, ?H f i.
Convexly combining together the two smoothness functionals (4) and (10), we obtain a
smoothness measure of function f defined on the whole vertex set V :
?? (f ) = ??A (f ) + (1 ? ?)?H (f ), 0 ? ? ? 1,
(11)
where the parameter ? weighs the relative importance between ?A (f ) and ?H (f ). Extend
the operator T to H(V ) by defining (T f )(v) = 0 if v is only in the authority set A and
not in the hub set H. Similarly extend T ? by defining (T ? f )(v) = 0 if v is only in the
hub set H and not in the authority set A. Then, if the remaining operators are extended
correspondingly, one can define the operator S? : H(V ) ? H(V ) by
S? = ?SA + (1 ? ?)SH ,
(12)
and the Laplacian on directed graphs ?? : H(V ) ? H(V ) by
?? = I ? S ? .
(13)
Clearly, ?? = ??A + (1 ? ?)?H . By Proposition 1 and 2, it is easy to see that:
Proposition 3. ?? (f ) = hf, ?? f i.
2.4
Regularization
Define a function y in H(V ) in which y(v) = 1 or ?1 if vertex v is labeled as positive
or negative, and 0, if it is not labeled. The classification problem can be regarded as the
problem of finding a function f , which reproduces the target function y to a sufficient
degree of accuracy while being smooth in a sense quantified by the above smoothness
functional. A formalization of this idea leads to the following optimization problem:
o
n
?
(14)
f ? = argmin ?? (f ) + kf ? yk2 .
2
f ?H(V )
The final classification of vertex v is obtained as sign f ? (v). The first term in the bracket
is called the smoothness term or regularizer, which measures the smoothness of function
f, and the second term is called the fitting term, which measures its closeness to the given
function y. The trade-off between these two competitive terms is captured by a positive
parameter ?. Successively smoother solutions f ? can be obtained by decreasing ? ? 0.
Theorem 4. The solution f ? of the optimization problem (14) satisfies
?? f ? + ?(f ? ? y) = 0.
Proof. By Proposition 3, we have
??? (f )
.
?f v
Differentiating the cost function in the bracket of (14) with respect to function f completes
the proof.
(?? f )(v) =
Corollary 5. The solution f ? of the optimization problem (14) is
f ? = (1 ? ?)(I ? ?S? )?1 y, where ? = 1/(1 + ?).
It is worth noting that the closed form solution presented by Corollary 5 shares the same
appearance as the algorithm proposed by [17], which operates on undirected graphs.
3
Experiments
We considered the Web page categorization task on the WebKB dataset [6]. We only
addressed a subset which contains the pages from the four universities: Cornell, Texas,
Washington, Wisconsin. We removed pages without incoming or outgoing links, resulting in 858, 825, 1195 and 1238 pages respectively, for a total of 4116. These pages were
manually classified into the following seven categories: student, faculty, staff, department,
course, project and other. We investigated two different classification tasks. The first is
used to illustrate the significance of connectivity information in classification, whereas the
second one stresses the importance of preserving the directionality of edges. We may assign a weight to each hyperlink according to the textual content of web pages or the anchor
text contained in hyperlinks. However, here we are only interested in how much we can
obtain from link structure only and hence adopt the canonical weight function defined in
Section 2.1.
We first study an extreme classification problem: predicting which university the pages
belong to from very few labeled training examples. Since pages within a university are
well-linked, and cross links between different universities are rare, we can imagine that
few training labels are enough to exactly classify pages based on link information only. For
each of the universities, we in turn viewed corresponding pages as positive examples and
the pages from the remaining universities as negative examples. We have randomly draw
two pages as the training examples under the constraint that there is at least one labeled
instance for each class. Parameters were set to ? = 0.50, and ? = 0.95. (In fact, in
this experiment the tuning parameters have almost no influence on the result.) Since the
Web graph is not connected, some small isolated subgraphs possibly do not contain labeled
instances. The values of our classifying function on the pages contained in these subgraphs
will be zeros and we simply think of these pages as negative examples. This is consistent
with the search engine ranking techniques [2, 13]. We compare our method with SVMs
using a kernel matrix K constructed as K = W T W [11], where W denotes the adjacency
matrix of the web graph and W T denotes the transpose of W . The test errors averaged over
100 training sample sets for both our method and SVMs are summarized into the following
table:
our method
SVMs
Cornel
0.03 (? 0.00)
0.42 (? 0.03)
Texas
0.02 (? 0.01)
0.39 (? 0.03)
Washington
0.01 (? 0.00)
0.40 (? 0.02)
Wisconsin
0.02 (? 0.00)
0.43 (? 0.02)
However, to be fair, we should state that the kernel matrix that we used in the SVM may not
be the best possible kernel matrix for this task ? this is an ongoing research issue which is
not the topic of the present paper.
The other investigated task is to discriminate the student pages in a university from the
non-student pages in the same university. As a baseline we have applied our regularization
method on the undirected graph [17] obtained by treating links as undirected or bidirectional, i.e., the affinity matrix is defined to be W T + W . We use the AUC scores to
measure the performances of the algorithms. The experimental results in Figure 3(a)-3(d)
clearly demonstrate that taking the directionality of edges into account can yield substantial accuracy gains. In addition, we also studied the influence of different choices for the
parameters ? and ?; we used the Cornell Web for that purpose and sampled 10 labeled
training pages. Figure 3(e) show that relatively small values of ? are more suitable. We
think that is because the subgraphs in each university are quite small, limiting the information conveyed in the graph structure. The influence of ? is shown in Figure 3(f). The
performance curve shows that large values for ? are preferable. This confirms the conjecture that co-link structure among authority pages is much more important than within the
hub set.
0.8
0.8
0.7
0.7
0.7
0.6
0.6
0.6
directed (?=1, ?=0.10)
undirected (?=0.10)
2
4
6
8
10
12
14
# labeled points
16
18
directed (?=1, ?=0.10)
undirected (?=0.10)
0.4
20
2
4
6
(a) Cornell
8
10
12
14
16
# labeled points
18
2
0.8
0.84
0.75
0.82
0.8
0.5
0.78
0.4
0.76
directed (?=1, ?=0.10)
undirected (?=0.10)
4
6
8
10
12
14
# labeled points
(d) Wisconsin
16
18
20
8
10
12
14
# labeled points
16
18
20
0.7
0.65
0.6
0.55
0.5
0.74
0.1
6
0.85
0.86
AUC
0.6
AUC
AUC
0.7
4
(c) Washington
0.88
2
directed (?=1, ?=0.10)
undirected (?=0.10)
0.3
20
(b) Texas
0.8
0.3
0.5
0.4
0.5
0.5
0.4
AUC
0.9
0.8
AUC
AUC
0.9
0.2
0.3
0.4
0.5
0.6
0.7
parameter values
(e) ? (Cornell)
0.8
0.9
0.45
0
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
parameter values
1
(f) ? (Cornell)
Figure 3: Classification on the WebKB dataset. Figure (a)-(d) depict the AUC scores of
the directed and undirected regularization methods on the classification problem student
vs. non-student in each university. Figure (e)-(f) illustrate the influences of the different
choices of the parameters ? and ?.
4
Conclusions
We proposed a general regularization framework on directed graphs, which has been validated on a real-word Web data set. The remaining problem is how to choose the suitable
parameters contained in this approach. In addition, it is worth noticing that this framework
can be applied without any essential changes to bipartite graphs, e.g. to graphs describing
customers? purchase behavior in market basket analysis. Moreover, in the absence of labeled instances, this framework can be utilized in an unsupervised setting as a (spectral)
clustering method for directed or bipartite graphs. Due to lack of space, we have not been
able to give a thorough discussion of these topics.
Acknowledgments We would like to thank David Gondek for his help on this work.
References
[1] M. Belkin, I. Matveeva, and P. Niyogi. Regularization and regression on large graphs. In COLT,
2004.
[2] S. Brin and L. Page. The anatomy of a large scale hypertextual web search engine. In Proc. 7th
Intl. WWW Conf., 1998.
[3] A. Broder, R. Kumar, F. Maghoul, P. Raghavan, S. Rajagopalan, R. Stata, A. Tomkins, and
J. Wiener. Graph structure in the Web. In Proc. 9th Intl. WWW Conf., 2000.
[4] S. Chakrabarti, B. Dom, and P. Indyk. Enhanced hypertext categorization using hyperlinks. In
Proc. ACM SIGMOD Conf., 1998.
[5] F. Chung. Spectral Graph Theory. Number 92 in Regional Conference Series in Mathematics.
American Mathematical Society, 1997.
[6] M. Craven, D. DiPasquo, D. Freitag, A. McCallum, T. Mitchell, K. Nigam, and S. Slattery.
Learning to extract symbolic knowledge from the World Wide Web. In Proc. 15th National
Conf. on Artificial Intelligence, 1998.
[7] J. Dean and M. Henzinger. Finding related Web pages in the World Wide Web. In Proc. 8th
Intl. WWW Conf., 1999.
[8] C. Ding, X. He, P. Husbands, H. Zha, and H. D. Simon. PageRank, HITS and a unified framework for link analysis. In Proc. 25th ACM SIGIR Conf., 2001.
[9] G. Flake, S. Lawrence, C. L. Giles, and F. Coetzee. Self-organization and identification of Web
communities. IEEE Computer, 35(3):66?71, 2002.
[10] C. Lee Giles, K. Bollacker, and S. Lawrence. CiteSeer: An automatic citation indexing system.
In Proc. 3rd ACM Conf. on Digital Libraries, 1998.
[11] T. Joachims, N. Cristianini, and J. Shawe-Taylor. Composite kernels for hypertext categorisation. In ICML, 2001.
[12] T. Joachims. Transductive learning via spectral graph partitioning. In ICML, 2003.
[13] J. Kleinberg. Authoritative sources in a hyperlinked environment.
46(5):604?632, 1999.
Journal of the ACM,
[14] R. Lempel and S. Moran. SALSA: the stochastic approach for link-structure analysis. ACM
Transactions on Information Systems, 19(2):131?160, 2001.
[15] A. Smola and R. Kondor. Kernels and regularization on graphs. In Learning Theory and Kernel
Machines. Springer-Verlag, Berlin-Heidelberg, 2003.
[16] V. N. Vapnik. Statistical learning theory. Wiley, NY, 1998.
[17] D. Zhou, O. Bousquet, T. N. Lal, J. Weston, and B. Sch o? lkopf. Learning with local and global
consistency. In NIPS, 2003.
[18] X. Zhu, Z. Ghahramani, and J. Lafferty. Semi-supervised learning using Gaussian fields and
harmonic functions. In ICML, 2003.
A
Proof of Proposition 1
Expand the right site of Equ. (4):
?A (f )
=
=
f (u)f (v)
f 2 (u)
?p
p(u)
p(u)p(v)
u,v h
2
X XX
f (u)f (v)
f (u) X X
?
ch ([u, v]) p
.
ch ([u, v])
p(u)
p(u)p(v)
u,v
u
v
XX
ch ([u, v])
(15)
h
h
By substituting Equ. (3), the first term in the above equality can be rewritten as
X X X w([h, u])w([h, v]) f 2 (u)
q(h)
p(u)
u
v
h
X X w([h, u]) 2
X 2
=
f (u) =
f (u).
p(u)
u
u
(16)
h
In addition, the second term in Equ. (15) can be transformed into
X X w([h, u])w([h, v]) f (u)f (v)
p
q(h)
p(u)p(v)
u,v h
XX
w([h, u]) w([h, v])
p
=
f (u) p
f (v).
q(h)p(u) q(h)p(v)
u,v h
Substituting Equ. (16) and (17) into (15), we have
X 2
XX
w([h, u]) w([h, v])
p
?A (f ) =
f (u) ?
f (u) p
f (v).
q(h)p(u) q(h)p(v)
u
u,v h
This completes the proof.
(17)
(18)
| 2718 |@word kondor:1 faculty:1 seems:2 norm:1 glue:1 confirms:1 citeseer:1 contains:1 score:2 series:1 document:1 comparing:1 yet:2 hofmann:1 treating:1 depict:1 v:1 intelligence:1 fewer:1 accordingly:2 mccallum:1 detecting:1 authority:21 node:19 mathematical:1 constructed:2 direct:1 chakrabarti:1 consists:4 freitag:1 fitting:1 pairwise:1 market:1 behavior:1 mpg:1 decreasing:1 encouraging:1 provided:1 xx:6 moreover:2 webkb:2 panel:4 project:1 kind:1 argmin:1 unified:1 finding:3 thorough:1 exactly:1 preferable:1 demonstrates:1 classifier:1 gondek:1 hit:1 partitioning:1 appear:1 planck:1 positive:5 before:2 understood:1 local:3 studied:1 quantified:1 co:12 averaged:1 directed:25 acknowledgment:1 definite:1 thought:1 composite:1 word:1 symbolic:1 unlabeled:3 operator:12 context:1 influence:4 www:3 dean:1 customer:1 sigir:1 assigns:1 subgraphs:4 utilizing:1 regarded:2 his:1 laplace:1 limiting:1 construction:3 target:1 imagine:1 enhanced:1 designing:1 matveeva:1 satisfying:1 utilized:1 labeled:13 ding:1 hypertext:2 schoelkopf:1 connected:1 trade:1 removed:1 substantial:1 intuition:1 environment:1 slattery:1 cristianini:1 dom:1 depend:1 bipartite:12 completely:1 represented:1 regularizer:1 distinct:2 artificial:1 quite:1 widely:1 niyogi:1 gi:1 think:4 transductive:1 final:1 indyk:1 eigenvalue:2 hyperlinked:1 propose:1 product:2 relevant:3 loop:1 combining:1 subgraph:1 adjoint:1 validate:1 olkopf:1 exploiting:1 parent:3 intl:3 categorization:5 help:1 illustrate:2 dengyong:2 measured:4 sa:4 c:1 come:1 anatomy:1 closely:1 attribute:1 stochastic:1 raghavan:1 brin:1 adjacency:1 premise:1 assign:1 preliminary:1 proposition:7 biological:1 considered:3 lawrence:2 pointing:1 substituting:2 major:1 adopt:2 purpose:1 proc:7 label:2 combinatorial:1 weighted:3 clearly:2 gaussian:1 rather:2 zhou:3 cornell:5 corollary:2 derived:2 focus:2 validated:1 joachim:2 likelihood:1 mainly:1 contrast:1 baseline:1 sense:1 typically:1 expand:1 transformed:1 interested:1 germany:1 issue:1 classification:13 among:2 colt:1 denoted:4 priori:1 yahoo:3 special:1 field:1 construct:1 having:1 washington:3 manually:1 biology:1 represents:2 unsupervised:1 icml:3 purchase:1 few:2 distinguishes:1 belkin:1 randomly:1 simultaneously:3 densely:1 national:1 individual:1 connects:1 organization:1 interest:2 investigate:1 highly:1 stata:1 evaluation:1 sh:3 bracket:2 extreme:1 behind:1 edge:8 indexed:1 taylor:1 penalizes:1 isolated:1 weighs:1 witnessed:1 column:1 classify:1 instance:3 giles:2 dipasquo:1 cost:1 vertex:24 subset:1 rare:1 successful:1 providence:1 cited:1 fundamental:1 broder:1 sequel:1 lee:1 off:1 connecting:1 together:3 connectivity:4 successively:1 choose:1 slowly:1 possibly:1 conf:7 book:1 american:1 chung:1 leading:1 account:1 suggestively:1 de:1 student:5 summarized:1 ranking:3 closed:1 linked:5 competitive:1 hf:5 zha:1 simon:1 accuracy:2 wiener:1 correspond:1 yield:1 lkopf:1 identification:1 worth:2 cybernetics:1 classified:1 husband:1 basket:1 henzinger:1 naturally:1 associated:2 proof:5 coetzee:1 gain:1 sampled:1 dataset:2 popular:1 mitchell:1 knowledge:1 bidirectional:1 supervised:3 though:1 strongly:1 smola:1 hand:1 web:36 overlapping:1 lack:1 google:1 usa:1 brown:2 concept:2 contain:2 normalized:2 regularization:10 hence:1 equality:1 deal:1 attractive:1 self:2 auc:8 stress:1 demonstrate:1 harmonic:1 common:5 functional:3 belong:2 extend:2 he:1 significant:2 smoothness:9 tuning:1 automatic:1 rd:1 mathematics:1 similarly:2 pointed:2 consistency:1 shawe:1 similarity:1 yk2:1 salsa:1 verlag:1 refrain:1 captured:1 preserving:1 staff:1 semi:4 smoother:1 infer:1 smooth:1 cross:1 retrieval:1 laplacian:4 regression:1 represent:1 kernel:13 addition:4 whereas:1 addressed:1 completes:2 source:1 sch:2 regional:1 induced:3 undirected:10 lafferty:1 call:1 noting:1 easy:1 enough:1 inner:2 idea:2 sibling:1 texas:3 linkage:4 svms:4 category:4 exist:1 canonical:2 notice:4 sign:1 arising:1 diverse:1 express:1 four:1 graph:54 concreteness:1 noticing:1 powerful:1 almost:1 draw:1 appendix:1 hypertextual:1 strength:2 constraint:1 categorisation:1 ri:1 sake:1 bousquet:1 kleinberg:3 kumar:1 relatively:1 conjecture:1 department:2 developing:1 according:2 craven:1 belonging:1 flake:1 slightly:1 indexing:1 computationally:2 mutually:1 turn:1 describing:1 available:1 endowed:1 rewritten:1 spectral:3 lempel:1 original:1 thomas:1 denotes:4 remaining:4 cf:2 clustering:1 tomkins:1 exploit:1 sigmod:1 ghahramani:1 establish:1 society:1 question:2 usual:1 exhibit:1 affinity:1 link:16 thank:1 berlin:1 topic:7 seven:1 extent:1 tuebingen:2 equip:1 relationship:3 innovation:1 statement:1 negative:4 inevitably:1 defining:3 situation:1 extended:1 community:2 introduced:1 david:1 namely:1 pair:1 connection:3 lal:1 engine:3 textual:1 nip:1 able:1 suggested:1 hyperlink:7 rajagopalan:1 pagerank:1 max:1 including:1 explanation:1 suitable:3 natural:1 force:1 predicting:1 zhu:1 representing:1 library:1 extract:1 text:1 kf:2 relative:1 wisconsin:3 interesting:1 digital:1 authoritative:2 degree:5 conveyed:1 sufficient:1 consistent:1 article:5 classifying:3 share:2 course:1 transpose:1 allow:1 institute:1 fall:1 wide:2 taking:1 correspondingly:1 differentiating:1 curve:1 world:4 ignores:1 reinforcement:1 transaction:1 functionals:2 citation:5 implicitly:1 bernhard:2 global:4 reproduces:1 incoming:1 anchor:1 equ:4 search:3 why:1 table:1 promising:1 composing:1 ca:2 nigam:1 heidelberg:1 investigated:2 complex:1 constructing:1 domain:1 significance:1 whole:1 child:1 fair:1 site:4 depicts:1 scattered:1 ny:1 wiley:1 formalization:1 theorem:1 specific:1 hub:18 moran:1 svm:1 convexly:1 closeness:1 essential:1 vapnik:1 importance:2 portal:1 flavor:1 simply:1 likely:4 appearance:1 ordered:1 contained:3 recommendation:1 springer:1 ch:6 cite:1 bollacker:1 satisfies:1 acm:5 weston:1 identity:1 viewed:1 absence:1 content:1 change:2 directionality:3 typical:2 operates:1 called:3 total:1 discriminate:1 experimental:2 support:1 relevance:3 ongoing:1 outgoing:1 |
1,892 | 2,719 | On Semi-Supervised Classification
Balaji Krishnapuram, David Williams, Ya Xue, Alex Hartemink, Lawrence Carin
Duke University, USA
M?ario A. T. Figueiredo
Instituto de Telecomunicac?o? es, Instituto Superior T?ecnico, Portugal
Abstract
A graph-based prior is proposed for parametric semi-supervised classification. The prior utilizes both labelled and unlabelled data; it also integrates features from multiple views of a given sample (e.g., multiple
sensors), thus implementing a Bayesian form of co-training. An EM
algorithm for training the classifier automatically adjusts the tradeoff between the contributions of: (a) the labelled data; (b) the unlabelled data;
and (c) the co-training information. Active label query selection is performed using a mutual information based criterion that explicitly uses the
unlabelled data and the co-training information. Encouraging results are
presented on public benchmarks and on measured data from single and
multiple sensors.
1
Introduction
In many pattern classification problems, the acquisition of labelled training data is costly
and/or time consuming, whereas unlabelled samples can be obtained easily. Semisupervised algorithms that learn from both labelled and unlabelled samples have been the
focus of much research in the last few years; a comprehensive review up to 2001 can be
found in [13], while more recent references include [1, 2, 6, 7, 16?18].
Most recent semi-supervised learning algorithms work by formulating the assumption that
?nearby? points, and points in the same structure (e.g., cluster), should have similar labels
[6,7,16]. This can be seen as a form of regularization, pushing the class boundaries toward
regions of low data density. This regularization is often implemented by associating the
vertices of a graph to all the (labelled and unlabelled) samples, and then formulating the
problem on the vertices of the graph [6, 16?18].
While current graph-based algorithms are inherently transductive ? i.e., they cannot be
used directly to classify samples not present when training ? our classifier is parametric and the learned classifier can be used directly on new samples. Furthermore, our algorithm is trained discriminatively by maximizing a concave objective function; thus we
avoid thorny local maxima issues that plague many earlier methods.
Unlike existing methods, our algorithm automatically learns the relative importance of the
labelled and unlabelled data. When multiple views of the same sample are provided (e.g.
features from different sensors), we develop a new Bayesian form of co-training [4]. In
addition, we also show how to exploit the unlabelled data and the redundant views of the
sample (from co-training) in order to improve active label query selection [15].
The paper is organized as follows. Sec. 2 briefly reviews multinomial logistic regression.
Sec. 3 describes the priors for semi-supervised learning and co-training. The EM algorithm
derived to learn the classifiers is presented in Sec. 4. Active label selection is discussed in
Sec. 5. Experimental results are shown in Sec. 6, followed by conclusions in Sec. 7.
2
Multinomial Logistic Regression
In an m-class supervised learning problem, one is given a labelled training set DL =
{(x1 , y 1 ), . . . , (xL , y L )}, where xi ? Rd is a feature vector and y i the corresponding
(1)
(m)
class label. In ?1-of-m? encoding, y i = [yi , . . . , yi ] is a binary vector, such that
(c)
(j)
yi = 1 and yi = 0, for j 6= c, indicates that sample i belongs to class c. In multinomial
logistic regression [5], the posterior class probabilities are modelled as
Pm
log P (y (c) = 1|x) = xT w(c) ? log k=1 exp(xT w(k) ), for c = 1, . . . , m, (1)
Pm
where w(c) ? Rd is the class-c weight vector. Notice that since c=1 P (y (c) = 1|x) = 1,
one of the weight vectors is redundant; we arbitrarily choose to set w(m) = 0, and consider
the (d (m?1))-dimensional vector w = [(w(1) )T , ..., (w(m?1) )T ]T . Estimation of w may
be achieved by maximizing the log-likelihood (with Y ? {y 1 , ..., y L }) [5]
Pm
PL Pm (c) T (c)
T (j)
y
x
w
?
log
`(w) ? log P (Y|w) = i=1
exp(x
w
)
. (2)
i
i
c=1 i
j=1
In the presence of a prior p(w), we seek a maximum a posteriori (MAP) estimate,
b = arg maxw {`(w) + log p(w)}. Actually, if the training data is separable, `(w) is
w
unbounded, and a prior is crucial.
Although we focus on linear classifiers, we may see the d-dimensional feature vectors x as
having resulted from some deterministic, maybe nonlinear, transformation of an input raw
feature vector r; e.g., in a kernel classifier, xi = [1, K(r i , r 1 ), ..., K(r i , r L )] (d = L + 1).
3
3.1
Graph-Based Data-Dependent Priors
Graph Laplacians and Regularization for Semi-Supervised Learning
Consider a scalar function f = [f1 , ..., f|V | ]T , defined on the set V = {1, 2, ..., |V |} of
vertices of an undirected graph (V, E). Each edge of the graph, joining vertices i and j, is
given a weight kij = kji ? 0, and we collect all the weights in a |V | ? |V | matrix K. A
natural way to measure how much f varies across the graph is by the quantity
XX
kij (fi ? fj )2 = 2 f T ? f ,
(3)
i
j
P
P
where ? = diag{ j k1j , ..., j k|V |j }?K is the so-called graph Laplacian [2]. Notice
that kij ? 0 (for all i, j) guarantees that ? is positive semi-definite and also that ? has (at
least) one null eigenvalue (1T?1 = 0, where 1 has all elements equal to one).
In semi-supervised learning, in addition to DL , we are given U unlabelled samples DU =
{xL+1 , . . . , xL+U }. To use (3) for semi-supervised learning, the usual choice is to assign
one vertex of the graph to each sample in X = [x1 , . . . , xL+U ]T (thus |V | = L + U ),
and to let kij represent some (non-negative) measure of ?similarity? between xi and xj . A
Gaussian random field (GRF) is defined on the vertices of V (with inverse variance ?)
p(f ) ? exp{?? f T? f /2},
in which configurations that vary more (according to (3)) are less probable. Most graphbased approaches estimate the values of f , given the labels, using p(f ) (or some modification thereof) as a prior. Accordingly, they work in a strictly transductive manner.
3.2
Non-Transductive Semi-Supervised Learning
We first consider two-class problems (m = 2, thus w ? Rd ). In contrast to previous uses of
graph-based priors, we define f as the real function f (defined over the entire observation
space) evaluated at the graph nodes. Specifically, f is defined as a linear function of x,
and at the graph node i, fi ? f (xi ) = wT xi . Then, f = [f1 , ..., f|V | ]T = Xw, and p(f )
induces a Gaussian prior on w, with precision matrix A = XT ?X,
p(w) ? exp{?(?/2) wT XT? Xw} = exp{?(?/2) wT Aw}.
(4)
Notice that since ? is singular, A may also be singular, and the corresponding prior may
therefore be improper. This is no problem for MAP estimation of w because (as is well
known) the normalization factor of the prior plays no role in this estimate. If we include
extra regularization, by adding a non-negative diagonal matrix to A, the prior becomes
p(w) ? exp ?(1/2) wT (?0 A + ?) w ,
(5)
where we may choose ? = diag{?1 , ..., ?d }, ? = ?1 I, or even ? = 0.
For m > 2, we define (m?1) identical independent priors, one for each w(c) , c = 1, ..., m.
The joint prior on w = [(w(1) )T , ..., (w(m?1) )T ]T is then
m?1
Y
1
1
(c)
p(w|?) ?
exp{? (w(c) )T ?0 A + ?(c) w(c) } = exp{? wT ?(?)w}, (6)
2
2
c=1
(c)
(c)
(c)
where ? is a vector containing all the ?i parameters, ?(c) = diag{?1 , ..., ?d }, and
(1)
(m?1)
?(?) = diag{?0 , ..., ?0
} ? A + block-diag{?(1) , ..., ?(m?1) }.
(7)
Finally, since all the ??s are inverses of variances, the conjugate priors are Gamma [3]:
(c)
(c)
(c)
(c)
p(?0 |?0 , ?0 ) = Ga(?0 |?0 , ?0 ), and p(?i |?1 , ?1 ) = Ga(?i |?1 , ?1 ), for c =
1, ..., m ? 1 and i = 1, ..., d. Usually, ?0 , ?0 , ?1 , and ?1 are given small values indicating
diffuse priors. In the zero limit, we obtain scale-invariant (improper) Jeffreys hyper-priors.
Summarizing, our model for semi-supervised learning includes the log-likelihood (2), a
prior (6), and Gamma hyper-priors. In Section 4, we present a simple and computationally
efficient expectation-maximization (EM) algorithm for obtaining the MAP estimate of w.
3.3
Exploiting Features from Multiple Sensors: The Co-Training Prior
In some applications several sensors are available, each providing a different set of features.
For simplicity, we assume two sensors s ? {1, 2}, but everything discussed here is easily
(s)
extended to any number of sensors. Denote the features from sensor s, for sample i, as xi ,
and Ss as the set of sample indices for which we have features from sensor s (S1 ?S2 =
{1, ..., L + U }). Let O = S1 ? S2 be the indices for which both sensors are available, and
OU = O ? {L + 1, ..., L + U } the unlabelled subset of O.
By using the samples in S1 and S2 as two independent training sets, we may obtain two sepb 1 and w
b 2 ). However, we can coordinate the information from
arate classifiers (denoted w
b1
both sensors by using an idea known as co-training [4]: on the OU samples, classifiers w
2
b should agree as much as possible. Notice that, in a logistic regression framework,
and w
the disagreement between the two classifiers on the OU samples can be measured by
P
1 T (1)
2 T (2) 2
T
(8)
i?OU [(w ) xi ? (w ) xi ] = ? C ?,
P
where ? = [(w1 )T (w2 )T ]T and C = i?OU [(x1i )T (?x2i )T ]T [(x1i )T (?x2i )T ]. This
suggests the ?co-training prior? (where ?co is an inverse variance):
p(w1 , w2 ) = p(?) ? exp ?(?co /2) ? TC ? .
(9)
This Gaussian prior can be combined with two smoothness Gaussian priors on w1 and w2
(obtained as described in Section 3.2); this leads to a prior which is still Gaussian,
p(w1 , w2 ) = p(?) ? exp ?(1/2) ? T ?co C + block-diag{?1 , ?2 } ? ,
(10)
where ?1 and ?2 are the two graph-based precision matrices (see (7)) for w1 and w2 .
We can again adopt a Gamma hyper-prior for ?co . Under this prior, and with a logistic
regression likelihood as above, estimates of w1 and w2 can easily be found using minor
modifications to the EM algorithm described in Section 4. Computationally, this is only
slightly more expensive than separately training the two classifiers.
4
Learning Via EM
b we use the EM algorithm, with ? as missing data, which
To find the MAP estimate w,
is equivalent to integrating out ? from the full posterior before maximization [8]. For
simplicity, we will only describe the single sensor case (no co-training).
E-step: We compute the expected value of the complete log-posterior, given Y and the
b Q(w|w)
b ? E[log p(w, ?|Y)|w].
b Since
current parameter estimate w:
log p(w, ?|Y) = log p(Y|w) ? (1/2)wT ?(?)w + K,
(11)
(where K collects all terms independent of w) is linear w.r.t. all the ? parameters (see (6)
and (7)), we just have to plug their conditional expectations into (11):
b = log p(Y|w) ? (1/2)wT E[?(?)|w]
b w = `(w) ? (1/2)wT ?(w)
b w. (12)
Q(w|w)
We consider several different choices for the structure of the ? matrix. The necessary
expectations have well-known closed forms, due to the use of conjugate Gamma hyper(c)
priors [3]. For example, if the ?0 are m ? 1 free non-negative parameters, we have
(c)
(c)
b = (2 ?0 + d) [2 ?0 + (w
b (c) )T Aw
b (c) ]?1 .
?0 ? E[?0 |w]
(c)
for c = 1, ..., m ? 1. For ?0 = ?0 , we still have a simple closed-form expres(c)
b and the same is true for the ?i parameters, for i > 0. Finally,
sion for E[?0 |w],
b ? E[?(?)|w]
b results from replacing the ??s in (7) by the corresponding conditional
?(w)
expectations.
b the M-step reduces to a logistic regression problem with a
M-step: Given matrix ?(w),
quadratic regularizer, i.e., maximizing (12). To this end, we adopt the bound optimization
approach (see details in [5, 11]). Let B be a positive definite matrix such that ?B bounds
below (in the matrix sense) the Hessian of `(w), which is negative definite, and g(w) is
b
the gradient of `(w). Then, we have the following lower bound on Q(w|w):
T
T
T
b ? l(w)
b + (w ? w)
b g(w)
b ? [(w ? w)
b B(w ? w)
b + w ?(w)w]/2.
b
Q(w|w)
b new = (B + ?(w))
b ?1 (Bw
b + g(w)),
b is guaranteed
The maximizer of this lower bound, w
new
b |w)
b ? Q(w|
b w),
b and we thus obtain a monotonic gento increase the Q-function, Q(w
eralized EM algorithm [5, 11]. This (maybe costly) matrix inversion can be avoided by a
sequential approach where we only maximize w.r.t. one element of w at a time, preserving
the monotonicity of the procedure. The sequential algorithm visits one particular element
of w, say wu , and updates its estimate by maximizing the bound derived above, while
keeping all other variables fixed at their previous values. This leads to
b ? (?(w)
b w)
b u ] [(B + ?(w))
b uu ]?1 ,
w
bunew = w
bu + [gu (w)
(13)
and
=w
bv , for v 6= u. The total time required by a full sweep for all u = 1, ..., d is
O(md(L + d)); this may be much better than the O((dm)3 ) of the matrix inversion.
w
bvnew
5
Active Label Selection
If we are allowed to obtain the label for one of the unlabelled samples, the following question arises: which sample, if labelled, would provide the most information?
b provided by EM. Our approach uses a Laplace approximaConsider the MAP estimate w
b H?1 ), where H is the posterior precision matrix,
tion of the posterior p(w|Y) ' N (w|w,
i.e., the Hessian of minus the log-posterior H = ?2 (? log p(w|Y)). This approximation is
b
known to be accurate for logistic regression under a Gaussian prior [14]. By treating ?(w)
(the expectation of ?(?)) as deterministic, we obtain an evidence-type approximation [14]
P
T
T
b
b + L
H = ?2 [? log(p(Y|w)p(w|?(w)))]
= ?(w)
i=1 (diag{pi } ? pi pi ) ? xi xi ,
where pi is the (m ? 1)-dimensional vector computed from (1), the c-th element of which
indicates the probability that sample xi belongs to class c.
Now let x? ? DU be an unlabelled sample and y ? its label. Assume that the MAP estib remains unchanged after including y ? . In Sec. 7 we will discuss the merits and
mate w
shortcomings of this assumption, which is only strictly valid when L ? ?. Accepting it
implies that after labeling x? , and regardless of y ? , the posterior precision changes to
(14)
H0 = H + (diag{p? } ? p? pT? ) ? x? xT? .
Since the entropy of a Gaussian with precision H is (?1/2) log |H| (up to an additive
constant), the mutual information (MI) between y ? and w (i.e., the expected decrease in
entropy of w when y ? is observed) is I(w; y ? ) = (1/2) log {|H0 |/|H|}. Our criterion
is then: the best sample to label is the one that maximizes I(w; y ? ). Further insight into
I(w; y ? ) can be obtained in the binary case (where p is a scalar); here, the matrix identity
|H + p? (1 ? p? )x? xT? | = |H|(1 + p? (1 ? p? )xT? H?1 x? ) yields
I(w; y ? ) = (1/2) log(1 + p? (1 ? p? )xT? H?1 x? ).
(15)
This MI is larger when p? ? 0.5, i.e., for samples with uncertain classifications. On the
other hand, with p? fixed, I(w; y ? ) grows with xT? H?1 x? , i.e., it is large for samples with
high variance of the corresponding class probability estimate. Summarizing, (15) favors
samples with uncertain class labels and high uncertainty in the class probability estimate.
6
Experimental Results
We begin by presenting two-dimensional synthetic examples to visually illustrate our semisupervised classifier. Fig. 1 shows the utility of using unlabelled data to improve the deci-
Figure 1: Synthetic two-dimensional examples. (a) Comparison of the supervised logistic
linear classifier (boundary shown as dashed line) learned only from the labelled data (shown
in color) with the proposed semi-supervised classifier (boundary shown as solid line) which
also uses the unlabelled samples (shown as dots). (b) A RBF kernel classifier obtained by
our algorithm, using two labelled samples (shaded circles) and many unlabelled samples.
Figure 2: (a)-(c) Accuracy (on UCI datasets) of the proposed method, the supervised SVM,
and the other semi-supervised classifiers mentioned in the text; a subset of samples is labelled and the others are treated as unlabelled samples. In (d), a separate holdout set is used
to evaluate the accuracy of our method versus the amount of labelled and unlabelled data.
sion boundary in linear and non-linear (kernel) classifiers (see figure caption for details).
Next we show results with linear classifiers on three UCI benchmark datasets. Results
with nonlinear kernels are similar, and therefore omitted to save space. We compare our
method against state-of-the-art semi-supervised classifiers: the GRF method of [18], the
SGT method of [10], and the transductive SVM (TSVM) of [9]. For reference, we also
present results for a standard SVM. To avoid unduly helping our method, we always use
a k=5 nearest neighbors graph, though our algorithm is not very sensitive to k. To avoid
disadvantaging other methods that do depend on such parameters, we use their best settings.
Since these adjustments cannot be made in practice, the difference between our algorithm
and the others is under-represented. Each point on the plots in Fig. 2(a)-(c) is an average
of 20 trials: we randomly select 20 labelled sets which are used by every method. All
remaining samples are used as unlabelled by the semi-supervised algorithms.
Figs. 2(a)-(c) are transductive, in the sense that the unlabelled and test data are the same.
Our logistic GRF is non-transductive: after being trained, it may be applied to classify new
data without re-training. In Fig. 2(d) we present non-transductive results for the Ionosphere
data. Training took place using labelled and unlabelled data, and testing was performed on
200 new unseen samples. The results suggest that semi-supervised classifiers are most
relevant when the labelled set is small relative to the unlabelled set (as is often the case).
Our final set of results address co-training (Sec. 3.3) and active learning (Sec. 5), applied to
airborne sensing data for the detection of surface and subsurface land mines. Two sensors
were used: (1) a 70-band hyper-spectral electro-optic (EOIR) sensor; (2) an X-band synthetic aperture radar (SAR). A simple (energy) ?prescreener? detected potential targets; for
each of these, two feature vectors were extracted, of sizes 420 and 9, for the EOIR and SAR
sensors, respectively. 123 samples have features from the EOIR sensor alone, 398 from the
Figure 3: (a) Land mine detection ROC curves of classifiers designed using only hyperspectral (EOIR) features, only SAR features, and both. (b) Number of landmines detected
during the active querying process (dotted lines), for active training and random selection
(for the latter the bars reflect one standard deviation about the mean). ROC curves (solid)
are for the learned classifier as applied to the remaining samples.
SAR sensor alone, and 316 from both. This data will be made available upon request.
We first consider supervised and semi-supervised classification. For the purely supervised
case, a sparseness prior is used (as in [14]). In both cases a linear classifier is employed. For
the data for which only one sensor is available, 20% of it is labelled (selected randomly).
For the data for which both sensors are available, 80% is labelled (again selected randomly).
The results presented in Fig. 3(a) show that, in general, the semi-supervised classifiers
outperform the corresponding supervised ones, and the classifier learned from both sensors
is markedly superior to classifiers learned from either sensor alone.
In a second illustration, we use the active-learning algorithm (Sec. 5) to only acquire the
100 most informative labels. For comparison, we also show average results over 100 independent realizations for random label query selection (error bars indicate one standard
deviation). The results in Fig. 3(b) are plotted in two stages: first, mines and clutter are selected during the labeling process (dashed curves); then, the 100 labelled examples are used
to build the final semi-supervised classifier, for which the ROC curve is obtained using the
remaining unlabelled data (solid curves). Interestingly, the active-learning algorithm finds
almost half of the mines while querying for labels. Due to physical limitations of the sensors, the rate at which mines are detected drops precipitously after approximately 90 mines
are detected ? i.e., the remaining mines are poorly matched to the sensor physics.
7
7.1
Discussion
Principal Contributions
Semi-supervised vs. Transductive: Unlike most earlier methods, after the training stage our
algorithm can directly classify new samples without computationally expensive re-training.
Tradeoff between labelled and unlabelled data: Automatically addressing the inherent
tradeoff between their relative contributions, we have ensured that even a small amount
of labelled data does not get overlooked because of an abundance of unlabelled samples.
Bayesian co-training: Using the proposed prior, classifiers for all sensors are improved
using: (a) the label information provided on the other types of data, and (b) samples drawn
from the joint distribution of features from multiple sensors.
Active label acquisition: We explicitly account for the knowledge of the unlabelled data and
the co-training information while computing the well known mutual information criterion.
7.2
Quality of Assumptions and Empirically Observed Shortcomings
The assumption that the mode of the posterior distribution of the classifier remains unchanged after seeing an additional label is clearly not true at the beginning of the active
learning procedure. However, we have empirically found it a very good approximation after the active learning procedure has yielded as few as 15 labels. This assumption allows a
tremendous saving in the computational cost, since it helps us avoid repeated re-training of
classifiers in the active label acquisition process while evaluating candidate queries.
A disturbing fact that has been reported in the literature (e.g., in [12]) and that we have
confirmed (in unreported experiments) is that the error rate of the active query selection
increases slightly when the number of labelled samples grows beyond an optimal number.
We conjecture that this may be caused by keeping the hyper-prior parameters ?0 , ?0 , ?1 , ?1
fixed at the same value; in all of our experiments we have set them to 10?4 , corresponding
to an almost uninformative hyper-prior.
References
[1] M. Belkin, I. Matveeva, and P. Niyogi. Regularization and regression on large graphs. In Proc.
Computational Learning Theory ? COLT?04, Banff, Canada, 2004.
[2] M. Belkin and P. Niyogi. Using manifold structure for partially labelled classification. In NIPS
15, MIT Press, Cambridge, MA, 2003.
[3] J. Bernardo and A. Smith. Bayesian Theory. J. Wiley & Sons, Chichester, UK, 1994.
[4] A. Blum and T. Mitchell. Combining labelled and unlabelled data with co-training. In Proc.
Computational Learning Theory ? COLT?98, Madison, WI, 1998.
[5] D. B?ohning. Multinomial logistic regression algorithm. Annals Inst. Stat. Math., vol. 44,
pp. 197?200, 1992.
[6] O. Chapelle, J. Weston, and B. Sch?olkopf. Cluster kernels for semi-supervised learning. In
NIPS 15, MIT Press, Cambridge, MA, 2003.
[7] A. Corduneanu and T. Jaakkola. On Information regularization. In Proc. Uncertainty in Artificial Intelligence ? UAI?03, Acapulco, Mexico, 2003.
[8] M. Figueiredo. Adaptive sparseness using Jeffreys? prior. In NIPS 14, MIT Press, 2002.
[9] T. Joachims. Transductive inference for text classification using support vector machines. In
Int. Conf. Machine Learning ? ICML?99, 1999.
[10] T. Joachims. Transductive learning via spectral graph partitioning. In ICML?03, 2003.
[11] K. Lange, D. Hunter, and I. Yang. Optimization transfer using surrogate objective functions. J.
Computational and Graphical Statistics, vol. 9, pp. 1?59, 2000.
[12] G. Schohn and D. Cohn. Less is more: Active learning with support vector machines. Intern.
Conf. on Mach. Learn. ? ICML?00.
[13] M. Seeger. Learning with labelled and unlabelled data. Tech. Rep., Institute for Adaptive and
Neural Computation, University of Edinburgh, UK, 2001.
[14] M. Tipping. Sparse Bayesian learning and the relevance vector machine. J. Mach. Learn.
Research, vol. 1, pp. 211?244, 2001.
[15] S. Tong and D. Koller. Support vector machine active learning with applications to text classification. In J. Mach. Learn. Research, vol. 2, pp. 45?66, 2001.
[16] D. Zhou, O. Bousquet, T. Lal, J. Weston, and B. Sch?olkopf. Semi-supervised learning by
maximizing smoothness. J. of Mach. Learn. Research, 2004 (submitted).
[17] X. Zhu, J. Lafferty and Z. Ghahramani. Combining active learning and semi-supervised learning
using Gaussian fields and harmonic functions. In ICML?03 Workshop on The Continuum from
Labelled to Unlabelled Data in Mach. Learning, 2003.
[18] X. Zhu, J. Lafferty and Z. Ghahramani. Semi-supervised learning: From Gaussian fields to
Gaussian processes. Tech. Rep. CMU-CS-03-175, School of CS, CMU, 2003.
| 2719 |@word trial:1 briefly:1 inversion:2 seek:1 minus:1 solid:3 configuration:1 interestingly:1 existing:1 current:2 additive:1 informative:1 treating:1 plot:1 update:1 designed:1 drop:1 alone:3 half:1 selected:3 v:1 intelligence:1 accordingly:1 beginning:1 smith:1 accepting:1 math:1 node:2 banff:1 unbounded:1 manner:1 expected:2 automatically:3 encouraging:1 becomes:1 provided:3 xx:1 begin:1 matched:1 maximizes:1 null:1 transformation:1 guarantee:1 every:1 bernardo:1 concave:1 ensured:1 classifier:29 uk:2 partitioning:1 before:1 positive:2 local:1 limit:1 instituto:2 encoding:1 joining:1 mach:5 approximately:1 collect:2 suggests:1 shaded:1 co:18 testing:1 practice:1 block:2 definite:3 procedure:3 integrating:1 seeing:1 krishnapuram:1 suggest:1 get:1 cannot:2 ga:2 selection:7 equivalent:1 map:6 deterministic:2 missing:1 maximizing:5 williams:1 regardless:1 simplicity:2 adjusts:1 insight:1 coordinate:1 laplace:1 sar:4 annals:1 pt:1 play:1 target:1 caption:1 duke:1 us:4 matveeva:1 element:4 expensive:2 balaji:1 observed:2 role:1 region:1 improper:2 decrease:1 mentioned:1 mine:7 radar:1 trained:2 depend:1 purely:1 upon:1 gu:1 easily:3 joint:2 represented:1 regularizer:1 describe:1 shortcoming:2 query:5 detected:4 labeling:2 artificial:1 hyper:7 h0:2 sgt:1 larger:1 say:1 s:1 estib:1 favor:1 niyogi:2 statistic:1 unseen:1 schohn:1 transductive:10 final:2 eigenvalue:1 took:1 uci:2 relevant:1 realization:1 combining:2 poorly:1 grf:3 olkopf:2 exploiting:1 cluster:2 help:1 illustrate:1 develop:1 stat:1 measured:2 nearest:1 school:1 minor:1 implemented:1 c:2 implies:1 uu:1 indicate:1 public:1 implementing:1 everything:1 assign:1 f1:2 probable:1 acapulco:1 strictly:2 pl:1 helping:1 exp:10 visually:1 lawrence:1 vary:1 adopt:2 continuum:1 omitted:1 estimation:2 proc:3 integrates:1 label:19 sensitive:1 mit:3 clearly:1 sensor:25 gaussian:10 always:1 subsurface:1 avoid:4 zhou:1 sion:2 jaakkola:1 derived:2 focus:2 joachim:2 indicates:2 likelihood:3 tech:2 contrast:1 seeger:1 summarizing:2 sense:2 posteriori:1 inst:1 inference:1 dependent:1 entire:1 koller:1 issue:1 classification:8 arg:1 colt:2 denoted:1 art:1 mutual:3 equal:1 field:3 saving:1 having:1 identical:1 icml:4 carin:1 others:2 inherent:1 few:2 belkin:2 randomly:3 gamma:4 resulted:1 comprehensive:1 bw:1 deci:1 detection:2 chichester:1 accurate:1 edge:1 necessary:1 circle:1 re:3 plotted:1 uncertain:2 kij:4 classify:3 earlier:2 maximization:2 cost:1 vertex:6 subset:2 deviation:2 addressing:1 reported:1 varies:1 aw:2 xue:1 synthetic:3 combined:1 density:1 bu:1 physic:1 eralized:1 w1:6 again:2 reflect:1 containing:1 choose:2 conf:2 account:1 potential:1 de:1 sec:10 includes:1 int:1 explicitly:2 caused:1 performed:2 view:3 tion:1 closed:2 graphbased:1 tsvm:1 contribution:3 accuracy:2 variance:4 yield:1 modelled:1 bayesian:5 raw:1 hunter:1 confirmed:1 expres:1 submitted:1 against:1 telecomunicac:1 energy:1 acquisition:3 pp:4 thereof:1 dm:1 mi:2 holdout:1 mitchell:1 color:1 knowledge:1 organized:1 ou:5 actually:1 thorny:1 tipping:1 supervised:28 improved:1 evaluated:1 though:1 furthermore:1 just:1 stage:2 hand:1 replacing:1 cohn:1 nonlinear:2 maximizer:1 logistic:10 mode:1 quality:1 corduneanu:1 grows:2 semisupervised:2 usa:1 true:2 regularization:6 during:2 criterion:3 presenting:1 ecnico:1 complete:1 fj:1 harmonic:1 fi:2 superior:2 multinomial:4 physical:1 empirically:2 discussed:2 cambridge:2 smoothness:2 rd:3 pm:4 portugal:1 unreported:1 dot:1 chapelle:1 similarity:1 surface:1 posterior:8 recent:2 belongs:2 binary:2 arbitrarily:1 rep:2 yi:4 seen:1 preserving:1 additional:1 employed:1 maximize:1 redundant:2 dashed:2 semi:23 multiple:6 full:2 reduces:1 unlabelled:28 plug:1 visit:1 laplacian:1 regression:9 expectation:5 cmu:2 kernel:5 represent:1 normalization:1 achieved:1 whereas:1 addition:2 separately:1 uninformative:1 airborne:1 singular:2 crucial:1 sch:2 extra:1 w2:6 unlike:2 markedly:1 undirected:1 electro:1 lafferty:2 presence:1 yang:1 xj:1 associating:1 lange:1 idea:1 tradeoff:3 utility:1 hessian:2 maybe:2 amount:2 clutter:1 band:2 induces:1 outperform:1 notice:4 dotted:1 vol:4 ario:1 blum:1 drawn:1 graph:18 year:1 inverse:3 uncertainty:2 place:1 almost:2 wu:1 utilizes:1 kji:1 bound:5 followed:1 guaranteed:1 quadratic:1 yielded:1 bv:1 optic:1 alex:1 diffuse:1 nearby:1 bousquet:1 formulating:2 separable:1 conjecture:1 according:1 request:1 conjugate:2 describes:1 across:1 em:8 slightly:2 son:1 wi:1 modification:2 s1:3 jeffreys:2 invariant:1 computationally:3 agree:1 remains:2 discus:1 merit:1 end:1 available:5 spectral:2 disagreement:1 save:1 remaining:4 include:2 graphical:1 madison:1 xw:2 pushing:1 exploit:1 ghahramani:2 build:1 unchanged:2 sweep:1 objective:2 question:1 quantity:1 parametric:2 costly:2 usual:1 diagonal:1 md:1 surrogate:1 gradient:1 separate:1 manifold:1 toward:1 index:2 illustration:1 providing:1 acquire:1 mexico:1 negative:4 observation:1 datasets:2 benchmark:2 mate:1 extended:1 canada:1 overlooked:1 david:1 required:1 lal:1 plague:1 learned:5 unduly:1 tremendous:1 nip:3 address:1 beyond:1 bar:2 usually:1 pattern:1 below:1 laplacians:1 including:1 natural:1 treated:1 zhu:2 improve:2 x2i:2 text:3 prior:33 review:2 literature:1 relative:3 precipitously:1 discriminatively:1 limitation:1 querying:2 versus:1 pi:4 land:2 last:1 free:1 keeping:2 figueiredo:2 landmines:1 institute:1 neighbor:1 sparse:1 edinburgh:1 boundary:4 curve:5 valid:1 evaluating:1 made:2 disturbing:1 ohning:1 avoided:1 adaptive:2 aperture:1 monotonicity:1 active:17 uai:1 b1:1 consuming:1 xi:11 learn:6 transfer:1 inherently:1 obtaining:1 du:2 diag:8 s2:3 allowed:1 repeated:1 x1:2 fig:6 roc:3 wiley:1 tong:1 precision:5 xl:4 x1i:2 candidate:1 learns:1 abundance:1 xt:9 sensing:1 svm:3 ionosphere:1 evidence:1 dl:2 workshop:1 adding:1 sequential:2 importance:1 hyperspectral:1 sparseness:2 entropy:2 tc:1 intern:1 adjustment:1 hartemink:1 partially:1 scalar:2 maxw:1 monotonic:1 extracted:1 ma:2 weston:2 conditional:2 identity:1 rbf:1 labelled:25 change:1 specifically:1 wt:8 principal:1 called:1 total:1 e:1 ya:1 experimental:2 indicating:1 select:1 support:3 latter:1 arises:1 relevance:1 evaluate:1 |
1,893 | 272 | 36
Bialek, Rieke, van Steveninck and Warland
Reading a Neural Code
William Bialek, Fred Rieke, R. R. de Ruyter van Steveninck 1 and
David Warland
Department of Physics, and
Department of Molecular and Cell Biology
University of California at Berkeley
Berkeley, California 94720
ABSTRACT
Traditional methods of studying neural coding characterize the encoding of known stimuli in average neural responses. Organisms
face nearly the opposite task - decoding short segments of a spike
train to extract information about an unknown, time-varying stimulus. Here we present strategies for characterizing the neural code
from the point of view of the organism, culminating in algorithms
for real-time stimulus reconstruction based on a single sample of
the spike train. These methods are applied to the design and analysis of experiments on an identified movement-sensitive neuron in
the fly visual system. As far as we know this is the first instance in
which a direct "reading" of the neural code has been accomplished.
1
Introduction
Sensory systems receive information at extremely high rates, and much of this information must be processed in real time. To understand real-time signal processing
in biological systems we must understand the representation of this information in
neural spike trains. \Ve ask several questions in particular:
? Does a single neuron signal only the occurrence of particular stimulus '"features," or can the spike train represent a continuous time-varying input?
1 Rijksuniversiteit
Groningen, Postbus 30.001,9700 RB Groningen The Netherlands
Reading a Neural Code
? How much information is carried by the spike train of a single neuron?
? Is the reliability of the encoded signal limited by noise at the sensory input
or by noise and inefficiencies in the subsequent layers of neural processing?
? Is the neural code robust to errors in spike timing, or do realistic levels of
synaptic noise place significant limits on information transmission?
? Do simple analog computations on the encoded signals correspond to simple
manipulations of the spike trains?
Although neural coding has been studied for more than fifty years, clear experimental answers to these questions have been elusive (Perkel & Bullock, 1968; de Ruyter
van Steveninck & Bialek, 1988). Here we present a new approach to the characterization of the neural code which provides explicit and sometimes surprising answers
to these questions when applied to an identified movement-sensitive neuron in the
fly visual system.
We approach the study of spiking neurons from the point of view of the organism,
which, based only on the spike train, must estimate properties of an unknown timevarying stimulus. Specifically we try to solve the problem of decoding the spike train
to recover the stimulus in real time. As far as we know our work is the first instance
in which it has been possible to "read" the neural code in this literal sense. Once
we can read the code, we can address the questions posed above. In this paper we
focus on the code reading algorithm, briefly summarizing the results which follow.
2
Theoretical background
The traditional approach to the study of neural coding characterizes the encoding
process: For an arbitrary stimulus waveform s( r), what can we predict about the
spike train? This process is completely specified by the conditional probability
distribution P[{tdls(r)] of the spike arrival times {til conditional on the stimulus
s( r). In practice one cannot characterize this distribution in its entirety; most
experiments result in only the lowest moment - the firing rate as function of time
given the stimulus.
The classic experiments of Adrian and others established that, for static stimuli, the
resulting constant firing rate provides a measure of stimulus strength. This concept
is easily extended to any stimulus waveform which is characterized by constant
parameters, such as a single frequency or fixed amplitude sine wave. l\'luch of the
effort in studying the encoding of sensory signals in the nervous system thus reduces
to probing the relation between these stimulus parameters and the resulting firing
rate. Generalizations to time-varying firing rates, especially in response to periodic
signals, have also been explored.
The firing rate is a continuous function of time which measures the probability
per unit time that the cell will generate a spike. The rate is thus by definition
an average quantity; it is not a property of a single spike train. The rate can
be estimated, in principle, by averaging over a large ensemble of redundant cells,
37
38
Bialek, Rieke, van Steveninck and Warland
or by averaging responses of a single cell over repeated presentations of the same
stimulus. This latter approach dominates the experimental study of spiking neurons.
Measurements of firing rate rely on some form of redundancy - either the spatial
redundancy of identical cells or the temporal redundancy of repeated stimuli. It is
simply not clear that such redundancy exists in real sensory systems under natural
stimulus conditions. In the absence of redundancy a characterization of neural
responses in terms of firing rate is oflittle relevance to the signal processing problems
faced by the organism. To say that "information is coded in firing rates" is of no
use unless one can explain how the organism could estimate these firing rates by
observing the spike trains of its own neurons.
We believe that none of the existing approaches2 to neural coding addresses the basic
problem of real-time signal processing with neural spike trains: The organism must
extract information about continuously varying stimulus waveforms using only the
discrete sequences of spikes. Real-time signal processing with neural spike trains
thus involves some sort of interpolation between the spikes that allows the organism
to estimate a continuous function of time.
The most basic problem of real-time signal processing is to decode the spike train and
recover an estimate of the stimulus waveform itself. Clearly if we can accomplish
this task then we can begin to understand how spike trains can be manipulated to
perform more complex computations; we can also address the quantitative issues
outlined in the Introduction. Because of the need to interpolate between spikes,
such decoding is not a simple matter of inverting the conventional stimulus-response
(rate) relations. In fact it is not obvious a priori that true decoding is even possible.
One approach to the decoding problem is to construct models of the encoding
process, and proceed analytically to develop algorithms for decoding within the
context of the model (Bialek & Zee, 1990). Using the results of this approach we
can predict that linear filtering will, under some conditions, be an effective decoding
algorithm, and we can determine the form of the filter itself. In this paper we have
a more limited goal, namely to see if the class of decoding algorithms identified
by Bialek and Zee is applicable to a real neuron. To this end we will treat the
structure of the decoding filter as unknown, and find the "best" filter under given
experimental conditions.
We imagine building a set of (generally non-linear) filters {Fn} which operate on
the spike train to produce an estimate of the stimulus. If the spikes arrive at times
{td, we write our estimate of the signal as a generalized convolution,
(1)
i
i,j
2Higher moments of the conditional probability P[{t i}ls(r)], such as the inter-spike interval
distribution (Perkel & Bullock, 1968) are also average properties, not properties of single spike
trains, and hence may not be relevant to real-time signal processing. White-noise methods (Marmarelis & Marmarelis, 1978) result in models which predict the time-varying firing rate in response
to arbitrary input waveforms and thus suffer the same limitations as other rate-based approaches.
Reading a Neural Code
How good are the reconstructions? We separate systematic and random errors by
introducing a frequency dependent gain g(w) such that (ls(w)l) g(w) (lsut(w)l).
The resulting gain is approximately unity through a reasonable bandwidth. Further,
the distribution of deviations between the stimulus and reconstruction is approxirr.ately Gaussian. The absence of systematic errors suggests that non-linearities
in the reconstruction filter are unlikely to help. Indeed, the contribution from the
st.. ~ond order term in Eq. (1) to the reconstructions is negligible.
=
---
o
~---------------------------,
~
"-
'"
~
--...
"0
":'
;...
"in
:3
~ ~
.::;
u
~
:ii'
~
.~
~
~
~
~----~____~__~____~____- J
o
10
frequency (Hz)
Figure 2: Spectral density of displacement noise from our reconstruction (upper
curve). By multiplying the displacement noise level by a bandwidth, we obtain the
square of the angular resolution of HI for a step displacement. For a reasonable
bandwidth the resolution is much less than the photoreceptor spacing, 1.350 "hyperacuity." Also shown is the limit to the resolution of small displacements set
by noise in the photoreceptor array (lower curve).
We identify the noise at frequency w as the difference between the stimulus and
the normalized reconstruction, n(w) = s(w) - g(w )Sed (w). \Ve then compute the
dpectral density (noise power per unit bandwidth) of the displacement noise (Fig 2).
The noise level achieved in HI is astonishing; with a one second integration time an
observer of the spike train in HI could judge the amplitude of a low frequency dither
to 0.01? - more than one hundred times less than the photoreceptor spacing! If the
fiY'f neural circuitry is noiseless, the fundamental limits to displacement resolution
39
40
Bialek, Rieke, van Steveninck and Warland
stimulus,
J
dw
Fl(T) =
.
_e- 1wr
27r
(s(w) Lj e- iwtj )
(Li,j eiW(t.-t ?)
(2)
?
j
The averages (.. . ) are with respect to an ensemble of stimuli
S(
T).
2. Minimize X 2 with respect to purely causal functions. This may be done analytically, or numerically by expanding F 1 ( T) in a complete set of functions
which vanish at negative times, then minimizing X2 by varying the coefficients
of the expansion. In this method we must explicitly introduce a delay time
which measures the lag between the true stimulus and our reconstruction.
We use the filter generated from the first method (which is the best possible linear
filter) to check the filter generated by the second method. Fig. 1 illustrates reconstructions using these two methods. The filters themselves are also shown in the
figure; we see that both methods give essentially the same answer.
...o
~ ~----------------------------~
~
~
-_N
'-'
::I
'"
I" I
I
\
,I' \
I
I
,
,,
I ,
I'
~ I'\
I
~
6b
-
~o
>.
'u0'1'
0
1:)
>
~
~
I
,,I
i"
\I
"v
,,Iv
Ull
III I
I
::I
t,~
::I
~
V~, .
i
...
:1
.
II
\
c,.o
,..
~
...v
J
.;::
I
I
'II I nil
~
d~
\V!
V
iii
-~
1
1 / 'I
'I
1\'
/.
~
'H II
I~,
r.
<Il
i
1\
,
If'
P
S'
:.;
1\"
I
I \! \
~
:;:
?
rJ~
\'
.
q
rf
0
,I
~II.
~
"
,~
I
~----------------------------~
I
I II
,
to:
~
II
0
~
'1'
2000
2100
2200
2300
time (msec)
2400
2500
?50
o
50
100
150
time (msec)
Figure 1: First order reconstruction se,,( T) using method 1 (solid line). The
st.imulus is shown here as a dotted line for comparison. The reconstruction shown
is for a segment of the spike train which was not used in the filter calculations. The
spike train is shown at the bottom of the figure, where the negative spikes are from
the "other eye" (cf. footnote 3). Both stimulus and reconstruction are smoothed
with a 5 msec half-wid th Gaussian filter. The filters calculating using both methods
are shown on the right.
Reading a Neural Code
We define the optimal filter to be that which minimizes X 2 = f dtls(t) - sest(t)1 2 ,
where s(t) is the true stimulus, and the integration is over the duration of the
experiment.
To insure that the filters we calculate allow real-time decoding, we require that the
filters be causal, for example FI(T < 0) = O. But the occurrence of a spike at
t' conveys information about the stimulus at a time t < t', so we must delay our
estimate of the stimulus by some time Tdelay > t' - t. In general we gain more
information by increasing the delay, so we face a tradeoff: Longer waiting times
allow us to gain more information but introduce longer reaction times to important
stimuli. This tradeoff is exactly the tradeoff faced by the organism in reacting to
external stimuli based on noisy and incomplete information.
3
Movelnent detection in the blowfly visual system
We apply our methods in experiments on a single wide field, movement-sensitive
neuron (H 1) in the visual system of the blowfly Calliphora erythrocephela. Flies
and other insects exhibit visually guided flight; during chasing behavior course
corrections can occur on time scales as short as 30 msec (Land & Collett, 1974). H1
appears to be an obligatory link in this control loop, encoding wide field horizontal
movements (Hausen, 1984). Given that the maximum firing rate in H1 is 100200 Hz, behavioral decisions must be based on the information carried by just a few
spikes from this neuron. Further, the horizontal motion detection system consists
of only a handful of neurons, so the fly has no opportunity to compute average
responses (or firing rates).
In the experiments described here, the fly is looking at a rigidly moving random pattern (de Ruyter van Steveninck, 1986). The pattern is presented on an oscilloscope,
and moved horizontally every 500 J-lsec in discrete steps chosen from an ensemble
which approximates Gaussian white noise. This time scale is short enough that we
can consider the resulting stimulus waveform s(t) to be the instantaneous angular
velocity. We record the spike arrival times {til extracellularly from the H1 neuron. 3
4
First order reconstructions
To reconstruct the stimulus waveform requires that we find the filter FI which
minimizes X2. We do this in two different ways:
1. Disregard the constraint that the filter be causal. In this case we can write
an explicit formula for the optimal filter in terms of the spike trains and the
3 There is one further caveat to the experiment. The firing rate in HI is increased for back-tofront motion and is decreased for front-to-back motion; the dynamic range is much greater in the
excitatory direction. The fiy, however, achieves high sensitivity in both directions by combining
information from both eyes. Because front-to-back motion in one eye corresponds to back-to-front
motion in the other eye, we can simulate the two eye case while recording from only one HI cell
by using an antisymmetric stimulus waveform. We combine the information coded in the spike
trains corresponding to the two "polarities" of the stimulus to obtain the information available
from both HI neurons.
41
42
Bialek, Rieke, van Steveninck and Warland
are set by noise in the photoreceptor array. We have calculated these limits in the
case where the displacements are small, which is true in our experiments at high
frequencies. In comparing these limits with the results in HI it is crucial that the
photoreceptor signal and noise characteristics (de Ruyter van Steveninck, 1986) are
measured under the same conditions as the HI experiments analyzed here. It is
clear from Fig. 2 that HI approaches the theoretical limit to its performance. We
emphasize that the noise spectrum in Fig. 2 is not a hypothetical measure of neural
performance. Rather it is the real noise level achieved in our reconstructions. As
far as we know this is the first instance in which the equivalent spectral noise level
of a spiking neuron has been measured.
To explore the tradeoff between the quality and delay of the reconstruction we
measure the cross-correlation of the smoothed stimulus with the reconstructions
calculated using method 2 above for delays of 10-70 msec. For a delay of 10 msec
the reconstruction carries essentially no information; this is expected since a delay of 10 msec is close to the intrinsic delay for phototransduction. As the delay
is increased the reconstructions improve, and this improvement saturates for delays greater than 40 msec, close to the behavioral reaction time of 30 msec - the
structure of the code is well matched to the behavioral decision task facing the
organism.
5
Conclusions
Learning how to read the neural code has allowed us to quantify the information
carried in the spike train independent of assumptions regarding the structure of
the code. In addition, our analysis gives some hopefully more general insights into
neural coding and computation:
1. The continuously varying movement signal encoded in the firing of H1 can be reconstructed by an astonishingly simple linear filter. If neurons summed their inputs
and marked the crossing of thresholds (as in many popular models), such reconstructions would be impossible; the threshold crossings are massively ambiguous
indicators of the signal waveform. We have carried out similar studies on a standard model neuron (the FitzHugh-Nagumo model), and find results similar to those
in the HI experiments. From the model neuron studies it appears that the linear
representation of signals in spike trains is a general property of neurons, at least in
a limited regime of their dynamics. In the near future we hope to investigate this
statement in other sensory systems.
2. The reconstruction is dominated by a "window" of - 40 msec during which
at most a few spikes are fired. Because so few spikes are important, it does not
make sense to talk about the "firing rate" - estimating the rate vs. time from
observations of the spike train is at least as hard as estimating the stimulus itself!
3. The quality of the reconstructions can be improved by accepting longer delays, but
this improvement saturates at - 30 - 40 msec, in good agreement with behavioral
decision times.
Reading a Neural Code
4. Having decoded the neural signal we obtain a meaningful estimate of the noise
level in the system and the information content of the code. H1 accomplishes a realtime version of hyper acuity, corresponding to a noise level near the limits imposed
by the quality of the sensory input. It appears that this system is close to achieving
optimal real-time signal processing.
5. From measurements of the fault tolerance of the code we can place requirements
on the noise levels in neural circuits using the information coded in H1. One of the
standard objections to discussions of "spike timing" as a mechanism of coding is
that there are no biologically plausible mechanisms which can make precise measurements of spike arrival times. We have tested the required timing precision by
introducing timing errors into the spike train and characterizing the resulting reconstructions. Remarkably the code is "fault tolerant," the reconstructions degrading
only slightly when we add timing errors of several msec.
Finally, we wish to emphasize our own surprise that it is so simple to recover time
dependent signals from neural spike trains. The filters we have constructed are not
very complicated, and they are linear. These results suggest that the representation
of time-dependent sensory data in the nervous system is much simpler than we migh t
have expected. We suggest that, correspondingly, simpler models of sensory signal
processing may be appropriate.
6
Acknowledglnents
We thank W. J. Bruno, M. Crair, L. Kruglyak, J. P. Miller, W. G. Owen, A. Zee,
and G. Zweig for many helpful discussions. This work was supported by the National Science Foundation through a Presidential Young Investigator Award to WB,
supplemented by funds from Cray Research and Sun Microsystems, and through a
Graduate Fellowship to FR. DW was supported in part by the System'S and Integrative Biology Training Program of the National Institutes of Health. Initial work was
supported by the Netherlands Organization for Pure Scientific Research (ZWO).
7
References
W. Bialek and A. Zee. J. Stat. Phys., in press, 1990.
K. Hausen. In M. Ali, editor, Photoreception and Vision in Invertebrates. Plenum
Press, New York and London, 1984.
M. Land and T. Collett. J. Compo Physiol., 89:331, 1974.
P. Marmarelis and V. Marmarelis. Analysis of Physiological Systems. The White
Noise Approach. Plenum Press, New York, 1978.
D. Perkel and T. Bullock. Neurosciences. Res. Prog. Bull., 6:221, 1968.
R. R. de Ruyter van Steveninck and W. Bialek. Proc. R. Soc. Lond. B, 234:379,
1988.
R. R. de Ruyter van Steveninck. Real-time Performance of a Movement-sensitive
Neuron in the Blowfly Visual System. Rijksuniversiteit Groningen, Groningen,
Netherlands, 1986.
43
| 272 |@word version:1 briefly:1 adrian:1 integrative:1 solid:1 carry:1 moment:2 initial:1 inefficiency:1 existing:1 reaction:2 comparing:1 surprising:1 must:7 fn:1 physiol:1 realistic:1 subsequent:1 fund:1 v:1 half:1 nervous:2 short:3 record:1 accepting:1 compo:1 caveat:1 characterization:2 provides:2 simpler:2 constructed:1 direct:1 consists:1 eiw:1 cray:1 combine:1 behavioral:4 introduce:2 inter:1 expected:2 indeed:1 behavior:1 themselves:1 oscilloscope:1 perkel:3 td:1 window:1 increasing:1 begin:1 estimating:2 linearity:1 insure:1 matched:1 circuit:1 lowest:1 what:1 minimizes:2 degrading:1 temporal:1 berkeley:2 quantitative:1 every:1 hypothetical:1 ull:1 exactly:1 control:1 unit:2 negligible:1 timing:5 treat:1 limit:7 encoding:5 rigidly:1 reacting:1 firing:15 interpolation:1 approximately:1 studied:1 fiy:2 suggests:1 limited:3 range:1 graduate:1 steveninck:10 ond:1 practice:1 chasing:1 displacement:7 suggest:2 cannot:1 close:3 context:1 impossible:1 conventional:1 equivalent:1 imposed:1 elusive:1 l:2 duration:1 resolution:4 pure:1 insight:1 array:2 dw:2 classic:1 rieke:5 plenum:2 imagine:1 decode:1 agreement:1 velocity:1 crossing:2 hyperacuity:1 bottom:1 fly:5 calculate:1 sun:1 movement:6 calliphora:1 dynamic:2 segment:2 astonishing:1 ali:1 purely:1 completely:1 easily:1 talk:1 train:27 effective:1 london:1 hyper:1 encoded:3 posed:1 solve:1 lag:1 say:1 plausible:1 imulus:1 reconstruct:1 presidential:1 itself:3 noisy:1 sequence:1 reconstruction:23 fr:1 relevant:1 loop:1 combining:1 fired:1 moved:1 transmission:1 requirement:1 produce:1 help:1 develop:1 stat:1 measured:2 eq:1 soc:1 entirety:1 involves:1 culminating:1 judge:1 quantify:1 direction:2 waveform:9 guided:1 filter:20 wid:1 require:1 generalization:1 biological:1 correction:1 visually:1 predict:3 circuitry:1 achieves:1 proc:1 applicable:1 sensitive:4 hope:1 clearly:1 gaussian:3 rather:1 varying:7 timevarying:1 groningen:4 focus:1 acuity:1 improvement:2 check:1 sense:2 summarizing:1 helpful:1 dependent:3 ately:1 unlikely:1 lj:1 relation:2 issue:1 insect:1 priori:1 spatial:1 integration:2 summed:1 field:2 once:1 construct:1 having:1 biology:2 identical:1 nearly:1 future:1 others:1 stimulus:37 few:3 manipulated:1 ve:2 interpolate:1 national:2 william:1 detection:2 organization:1 investigate:1 analyzed:1 zee:4 unless:1 iv:1 incomplete:1 re:1 causal:3 theoretical:2 rijksuniversiteit:2 instance:3 increased:2 wb:1 bull:1 introducing:2 deviation:1 hundred:1 delay:11 front:3 characterize:2 answer:3 periodic:1 accomplish:1 st:2 density:2 fundamental:1 sensitivity:1 systematic:2 physic:1 decoding:10 continuously:2 marmarelis:4 literal:1 external:1 til:2 li:1 de:6 kruglyak:1 coding:6 coefficient:1 matter:1 explicitly:1 sine:1 view:2 try:1 observer:1 h1:6 observing:1 characterizes:1 extracellularly:1 wave:1 sort:1 recover:3 complicated:1 contribution:1 sed:1 square:1 minimize:1 il:1 characteristic:1 ensemble:3 correspond:1 identify:1 miller:1 none:1 multiplying:1 explain:1 footnote:1 phys:1 synaptic:1 definition:1 frequency:6 obvious:1 conveys:1 static:1 gain:4 popular:1 ask:1 amplitude:2 back:4 appears:3 higher:1 follow:1 response:7 improved:1 done:1 angular:2 just:1 correlation:1 flight:1 horizontal:2 hopefully:1 quality:3 scientific:1 believe:1 building:1 concept:1 true:4 normalized:1 analytically:2 hence:1 read:3 white:3 during:2 ambiguous:1 generalized:1 complete:1 motion:5 instantaneous:1 fi:2 spiking:3 analog:1 organism:9 approximates:1 numerically:1 significant:1 measurement:3 outlined:1 phototransduction:1 bruno:1 reliability:1 sest:1 moving:1 longer:3 add:1 own:2 manipulation:1 massively:1 fault:2 accomplished:1 greater:2 accomplishes:1 determine:1 redundant:1 signal:20 ii:7 u0:1 rj:1 reduces:1 characterized:1 calculation:1 cross:1 zweig:1 nagumo:1 crair:1 molecular:1 award:1 coded:3 basic:2 noiseless:1 essentially:2 vision:1 represent:1 sometimes:1 achieved:2 cell:6 receive:1 background:1 addition:1 remarkably:1 spacing:2 interval:1 decreased:1 objection:1 fellowship:1 crucial:1 fifty:1 operate:1 hz:2 recording:1 collett:2 near:2 iii:2 enough:1 identified:3 opposite:1 bandwidth:4 regarding:1 tradeoff:4 effort:1 suffer:1 proceed:1 york:2 generally:1 clear:3 se:1 netherlands:3 processed:1 generate:1 dotted:1 estimated:1 neuroscience:1 per:2 rb:1 wr:1 discrete:2 write:2 waiting:1 redundancy:5 threshold:2 achieving:1 year:1 place:2 arrive:1 reasonable:2 prog:1 realtime:1 decision:3 layer:1 hi:10 fl:1 strength:1 occur:1 handful:1 constraint:1 x2:2 invertebrate:1 dominated:1 lsec:1 simulate:1 extremely:1 lond:1 fitzhugh:1 department:2 slightly:1 unity:1 bullock:3 biologically:1 mechanism:2 know:3 end:1 studying:2 available:1 apply:1 blowfly:3 spectral:2 appropriate:1 occurrence:2 cf:1 opportunity:1 calculating:1 warland:5 especially:1 question:4 quantity:1 spike:43 strategy:1 traditional:2 bialek:10 exhibit:1 separate:1 link:1 thank:1 code:18 polarity:1 minimizing:1 statement:1 negative:2 design:1 unknown:3 perform:1 upper:1 neuron:19 convolution:1 observation:1 extended:1 looking:1 saturates:2 precise:1 smoothed:2 arbitrary:2 david:1 inverting:1 namely:1 required:1 specified:1 california:2 established:1 address:3 microsystems:1 pattern:2 regime:1 reading:7 program:1 rf:1 power:1 natural:1 rely:1 indicator:1 improve:1 eye:5 carried:4 extract:2 health:1 faced:2 limitation:1 filtering:1 facing:1 foundation:1 astonishingly:1 principle:1 editor:1 tdelay:1 land:2 course:1 excitatory:1 supported:3 allow:2 understand:3 institute:1 wide:2 face:2 characterizing:2 correspondingly:1 van:10 tolerance:1 curve:2 calculated:2 fred:1 sensory:8 hausen:2 far:3 reconstructed:1 emphasize:2 tolerant:1 spectrum:1 continuous:3 ruyter:6 robust:1 expanding:1 expansion:1 complex:1 antisymmetric:1 noise:21 arrival:3 repeated:2 allowed:1 fig:4 probing:1 precision:1 decoded:1 explicit:2 msec:12 wish:1 vanish:1 young:1 formula:1 supplemented:1 explored:1 physiological:1 dominates:1 exists:1 intrinsic:1 illustrates:1 photoreception:1 surprise:1 simply:1 explore:1 visual:5 horizontally:1 corresponds:1 conditional:3 goal:1 presentation:1 marked:1 owen:1 absence:2 content:1 hard:1 specifically:1 averaging:2 nil:1 experimental:3 disregard:1 photoreceptor:5 meaningful:1 latter:1 relevance:1 investigator:1 tested:1 |
1,894 | 2,720 | Trait selection for assessing beef meat quality
using non-linear SVM
J.J. del Coz, G. F. Bay?on, J. D??ez,
O. Luaces, A. Bahamonde
Artificial Intelligence Center
University of Oviedo at Gij?on
[email protected]
?
Carlos Sanudo
Facultad de Veterinaria
University of Zaragoza
[email protected]
Abstract
In this paper we show that it is possible to model sensory impressions
of consumers about beef meat. This is not a straightforward task; the
reason is that when we are aiming to induce a function that maps object
descriptions into ratings, we must consider that consumers? ratings are
just a way to express their preferences about the products presented in
the same testing session. Therefore, we had to use a special purpose
SVM polynomial kernel. The training data set used collects the ratings of
panels of experts and consumers; the meat was provided by 103 bovines
of 7 Spanish breeds with different carcass weights and aging periods.
Additionally, to gain insight into consumer preferences, we used feature
subset selection tools. The result is that aging is the most important trait
for improving consumers? appreciation of beef meat.
1
Introduction
The quality of beef meat is appreciated through sensory impressions, and therefore its
assessment is very subjective. However, it is known that there are objective traits very important for the final properties of beef meat; this includes the breed and feeding of animals,
weight of carcasses, and aging of meat after slaughter. To discover the influence of these
and other attributes, we have applied Machine Learning tools to the results of an experience reported in [8]. In the experience, 103 bovines of 7 Spanish breeds were slaughtered
to obtain two kinds of carcasses, light and standard [5]; the meat was prepared with 3 aging
periods, 1, 7, and 21 days. Finally, the meat was consumed by a group, called panel, of 11
experts, and assessed by a panel of untrained consumers.
The conceptual framework used for the study reported in this paper was the analysis of
sensory data. In general, this kind of analysis is used for food industries in order to adapt
their productive processes to improve the acceptability of their specialties. They need to
discover the relationship between descriptions of their products and consumers? sensory
degree of satisfaction. An excellent survey of the use of sensory data analysis in the food
industry can be found in [15, 2]; for a Machine Learning perspective, see [3, 9, 6].
The role played by each panel, experts and consumers, is very clear. So, the experts? panel
is made up of a usually small group of trained people who rate several traits of products such
as fibrosis, flavor, odor, etc. . . The most essential property of expert panelists, in addition
to their discriminatory capacity, is their own coherence, but not necessarily the uniformity
of the group. Experts? panel can be viewed as a bundle of sophisticated sensors whose
ratings are used to describe each product, in addition to other objective traits. On the other
hand, the group of untrained consumers (C) are asked to rate their degree of acceptance or
satisfaction about the tested products on a given scale. Usually, this panel is organized in
a set of testing sessions, where a group of potential consumers assess some instances from
a sample E of the tested product. Frequently, each consumer only participates in a small
number (sometimes only one) of testing sessions, usually in the same day.
In general, the success of sensory analysis relies on the capability to identify, with a precise
description, a kind of product that should be reproducible as many times as we need to be
tested for as many consumers as possible. Therefore, the study of beef meat sensory quality
is very difficult. The main reason is that there are important individual differences in each
piece of meat, and the repeatability of tests can be only partially ensured. Notice that from
each animal there are only a limited amount of similar pieces of meat, and thus we can only
provide pieces of a given breed, weight, and aging period. Additionally, it is worthy noting
that the cost of acquisition of this kind of sensory data is very high.
The paper is organized as follows: in the next section we present an approach to deal with
testing sessions explicitly. The overall idea is to look for a preference or ranking function
able to reproduce the implicit ordering of products given by consumers instead of trying
to predict the exact value of consumer ratings; such function must return higher values
to those products with higher ratings. In Section 3 we show how some state of the art
FSS methods designed for SVM (Support Vector Machines) with non-linear kernels can be
adapted to preference learning. Finally, at the end of the paper, we return to the data set
of beef meat to show how it is possible to explain consumer behavior, and to interpret the
relevance of meat traits in this context.
2
Learning from sensory data
A straightforward approach to handle sensory data can be based on regression, where sensory descriptions of each object x ? E are endowed with the degree of satisfaction r(x)
for each consumer (or the average of a group of consumers). However, this approach does
not faithfully captures people?s preferences [7, 6]: consumers? ratings actually express a
relative ordering, so there is a kind of batch effect that often biases their ratings. Thus, a
product could obtain a higher (lower) rating depending on if it is assessed together with
worse (better) products. Therefore, information about batches tested by consumers in each
rating session is a very important issue. On the other hand, more traditional approaches,
such as testing some statistical hypotheses [16, 15, 2] require all available food products in
sample E to be assessed by the set of consumers C, a requisite very difficult to fulfill.
In this paper we use an approach to sensory data analysis based on learning consumers?
preferences, see [11, 14, 1], where training examples are represented by preference judgments, i.e. pairs of vectors (v, u) indicating that, for someone, object v is preferable to
object u. We will show that this approach can induce more useful knowledge than other
approaches, like regression based methods. The main reason is due to the fact that preference judgments sets can represent more relevant information to discover consumers? preferences.
2.1
A formal framework to learn consumer preferences
In order to learn our preference problems, we will try to find a real ranking function f that
maximizes the probability of having f (v) > f (u) whenever v is preferable to u [11, 14, 1].
Our input data is made up of a set of ratings (ri (x) : x ? Ei ) for i ? C. To avoid the batch
effect, we will create a preference judgment set P J = {v j > uj : j = 1, . . . , n} suitable
for our needs just considering all pairs (v, u) such that objects v and u were presented in
the same session to a given consumer i, and ri (v) > ri (u).
Thus, following the approach introduced in [11], we look for a function F : Rd ? Rd ? R
such that
?x, y ? Rd , F (x, y) > 0 ? F (x, 0) > F (y, 0).
(1)
Then, the ranking function f : Rd ? R can be simply defined by f (x) = F (x, 0).
As we have already constructed a set of preference judgments P J, we can specify F by
means of the restrictions
F (v j , uj ) > 0 and F (uj , v j ) < 0, ?j = 1, . . . , n.
(2)
Therefore, we have a binary classification problem that can be solved using SVM. We
follow the same steps as Herbrich et al. in [11], and define a kernel K as follows
K(x1 , x2 , x3 , x4 ) = k(x1 , x3 ) ? k(x1 , x4 ) ? k(x2 , x3 ) + k(x2 , x4 )
(3)
where k(x, y) = h?(x), ?(y)i is a kernel function defined as the inner product of two
objects represented in the feature space by their ? images. In the experiments reported in
Section 4, we will employ a polynomial kernel, defining k(x, y) = (hx, yi + c)g , with
c = 1 and g = 2. Notice that, finally we can express the ranking function f in a non-linear
form:
n
X
(1)
(2)
?i zi (k(xi , x) ? k(xi , x))
(4)
f (x) =
i=1
3
Feature subset selection methods in a non-linear environment
When dealing with sensory data, it is important to know not only which classifier is the
best and how accurate it is, but also which features are relevant for the tastes of consumers.
Producers can focus on these features to improve the quality of the final product. Additionaly, reductions on the number of features often lead to a cheaper data acquisition labour,
making these systems suitable for industrial operation [9].
There are many feature subset selection methods applied to SVM classification. If our
goal is to find a linear separator, RFE (Recursive Feature Elimination) [10] will be a good
choice. It is a ranking method that returns an ordering of the features. RFE iteratively
removes the less useful feature. This process is repeated until there are no more features.
Thus, we obtain an ordered sequence of features.
Following the main idea of RFE, we have used two methods capable of ordering features in
non-linear scenarios. We must also point that, in this case, preference learning data sets are
formed by pairs of objects (v, u), and each object in the pair has the same set of features.
Thus, we must modify the ranking methods so they can deal with the duplicated features.
3.1
Ranking features for non-linear preference learning
Method 1.- This method orders the list of features according to their influence in the
variations of the weights. It is a gradient-like method, introduced in [17], and found to be
a generalization of RFE to the non-linear case. It removes in each iteration the feature that
minimizes the ranking value
X
?K(s ? xk , s ? xj )
R1 (i) = |?i kwk2 | =
?k ?j zk zj
(5)
, i = 1, . . . , d
?si
k,j
where s is a scaling factor used to simplify the computation of partial derivatives. Due to
the fact that we are working on a preference learning problem, we need 4 copies of the
scaling factor. In this formula, for a polynomial kernel k(x, y) = (hx, yi + c)g and a
vector s such that ?i, si = 1 we have that
?k(s ? x, s ? y)
= 2g(xi yi )(c + hx, yi)g?1 .
(6)
?si
Method 2.- This method, introduced in [4], works in an iterative way; removing each
time the feature which minimizes the loss of predictive performance. When using this
method for preference learning with the kernel of equation (3) the ranking criterion can be
expressed as
?
?
X
X
(1),i
(2),i
(1),i
(2),i
R2 (i) = ?
zk ?
?j zj K(xj , xj , xk , xk )?
(7)
k
j
i
where x denotes a vector describing an object where the value for the i-th feature was
replaced by its mean value. Notice that a higher value of R2 (i), that is, a higher accuracy
on the training set when replacing feature i-th, means a lower relevance of that feature.
Therefore, we will remove the feature yielding the highest ranking value, as opposite to the
ranking method described previously.
3.2
Model selection on an ordered sequence of feature subsets
Once we have an ordering of the features, we must select the subset Fi which maximizes
the generalization performance of the system. The most common choice for a model selection method is cross-validation (CV), but its efficiency and high variance [1] lead us to
try another kind of methods. We have used ADJ (ADJusted distance estimate)[19]. This
is a metric-based method that selects one from a nested sequence of complexity-increasing
models. We construct a sequence of subsets F1 ? F2 ? . . . ? Fd , where Fi represents the
subset containing only the i most relevant features. Then we can create a nested sequence
of models fi , each one of these induced by SVM from the corresponding Fi .
The key idea is the definition of a metric on the space of hypothesis. Thus, given two
different hypothesis f and g, their distance is calculated as the expected disagreement in
their predictions. Given that these distances can only be approximated, ADJ establish a
? t), an adjusted distance estimate between any hypothesis f and
method to compute d(g,
the true target classification function t. Therefore, the selected hypothesis is
? l , t).
fk = arg min d(f
(8)
fl
? is computed by means of the expected disagreement in the
The estimation of distance, d,
predictions in a couple of sets: the training set T , and a set U of unlabeled examples, that is,
a set of cases sampled from the same distribution of T but for which the pretended correct
output is not given. The ADJ estimation is given by
dU (fk , fl )
def
(9)
ADJ(fl , t) = dT (fl , t) ? max
k<l dT (fk , fl )
where, for a given subset of examples S, dS (f, g) is the expected disagreement of hypothesis f and g in S. To avoid the impossibility of using the previous equation when there are
zero disagreements in T for two hypotheses we use the Laplace correction to the probability
estimation; thus,
!
X
1
def
dS (f, g) =
1+
1f (x)6=g(x)
(10)
|S| + 2
x?S
In general, it is not straightforward to obtain a set of unlabeled examples. However, for
learning preferences, we can easily build the set of unlabeled examples from a set of preference judgments formed by pairs of real objects randomly selected from the original preference judgment pairs.
3.3
Summarizing the data: dealing with redundancy
As we have previously pointed out, sensory data include ratings of experts for different
characteristics of the products, as well as physical and chemical features directly measured
on them. It is not infrequent to find out that some of these features are highly correlated;
some experts may have similar opinions about a certain feature, and similarities among
several chemical and physical features may be possible as well. In order to take advantage of these peculiarity, we have developed a simple redundancy filter, RF. It is meant
to be applied before any feature subset selection method, allowing us to discover intrinsic
redundancies in the data and to reduce the number of features used.
RF is an iterative process where each step gives rise to a new description of the original data
set. The two most similar features are replaced by a new one whose values are computed as
the average of them. Considering two given features ai and aj as (column) vectors whose
dimension is the number of examples in the data set, the similarity can be estimated by
means of their cosine, that is,
hai , aj i
similarity(ai , aj ) =
(11)
kai k ? kaj k
Applying this method we obtain a sequence of different descriptions of the original data set,
each one with one feature less than the previous. To select an adequate description in terms
of prediction accuracy, we use again ADJ. The selected description can be considered a
summarized version of the original data set to be processed by the feature subset selection
methods previously described.
4
Experimental results
In this section we show the experimental results obtained when we applied the tools described in previous sections to the beef meat data base [8]. Each piece of meat was described by 147 attributes: weight of the animal, breed (7 boolean attributes), aging, 6 physical attributes describing its texture and 12 sensory traits rated by 11 different experts (132
ratings). The meat comes from 103 bovines of 7 Spanish breeds (from 13 to 16 animals
of each breed); animals were slaughtered in order to obtain 54 light and 49 standard carcasses, uniformly distributed across breeds. In each rating session, 4 or 5 pieces of meat
were tested and a group of consumers were asked to rate only three different aspects: tenderness, flavor and acceptance. These three data sets have over 2420 preference judgments.
All the results shown in this section have been obtained by a 10-fold cross-validation.
4.1
Preference learning vs. regression
First, we performed a comparison between preference learning and regression methods.
We have experimented with a simple linear regression and with a well reputed regression
algorithm: Cubist, a commercial product from RuleQuest Research. To interpret regression
results we used the relative mean absolute deviation (Rmad), which is the quotient between
the mean absolute distance of the function learned by the regression method and the mean
absolute distance of the constant predictor that returns the mean value in all cases.
On the other hand, we can obtain some preference judgments from the ratings of the sessions comparing the rating of each product with the rest, one by one, and constructing the
Table 1: Regression and preference learning scores on beef meat data sets.
Regression
Preferences
Linear
Cubist Linear Cubist SVMl
SVMp
Error
Error
Error
Error
Rmad
Rmad
tenderness
flavor
acceptance
96.3%
99.3%
94.0%
97.8%
103.4%
97.2%
41.5%
43.8%
38.4%
43.1%
46.5%
40.2%
29.6%
32.7%
31.9%
19.4%
23.8%
22.1%
Avg.
96.5%
99.5%
41.2%
43.3%
31.4%
21.8%
corresponding pair. To learn from preference judgment data sets we used SVMlight [13]
with linear and polynomial kernels. In this case, the errors have a straightforward meaning as misclassifications; so in order to allow a fair comparison between regression and
preference learning approaches, we also tested regression models on preference judgments
test sets, calculating their misclassifications. The scores achieved on the three data sets
described previously, are shown in Table 1.
We can observe that regression methods are unable to learn any useful knowledge: their
relative mean absolute deviation (Rmad) is near 100% in all cases, that is, regression models usually perform equal than the constant predictor forecasting the mean value. From a
practical point of view, these results mean that raw consumers? ratings can not be used to
measure the overall sensory opinions. Even when these regression models are tested on
preference judgment sets the percentage of misclassifications is over 40%, clearly higher
than those obtained when using the preference learning approach. SVM-based methods can
reduce these errors down to an average near 30% with a linear kernel (SVMl ), and near 20%
if the kernel is a polynomial of degree 2 (SVMp ). This improvement shows that non-linear
kernels can explain consumers preferences better.
4.2
Feature selection
We used the FSS tools to gain insight into consumer preferences. For the sake of simplicity,
in what follows FSS1 and FSS2 will denote the feature subset selectors that use ranking
Method 1 and Method 2 respectively. The learner used in these experiments was SVMp
because it was the most accurate in previous tests (see Table 1). Given the size of the
three data sets it is almost impractical to use FSS1 and FSS2 due to its computational cost,
unless a previous reduction in the number of features can be achieved; therefore, in both
cases we used RF as a previous filter. Additionally, to improve the overall speed, features
were removed in chunks of five. In all cases we used ADJ to choose among the subsets of
features. We can see in Table 2 that FSS1 , FSS2 , and RF considerably reduce the number
of features at the expense of accuracy: it slightly decreases when we use the RF filter with
respect to the accuracy obtained on the original data set by SVMp ; it also decreases when
using FSS1 and FSS2 after RF.
The most useful result obtained from feature selection is the ranking list of traits. We
concentrate our study in tenderness and acceptance categories because they are more interesting from the point of view of beef meat producers. So, in acceptance data set, the three
most useful traits are: aging, breed and fibrosis. Some research works in the beef meat field
corroborate the importance of these characteristics [18, 12]. Specially, aging is crucial to
improve consumer acceptance. On the other hand, fibrosis is closely related with tenderness: the less fibrosis, the more tenderness. Usually many consumers identify tenderness
with acceptance, in the sense that a higher tenderness yields to a higher acceptance; then,
fibrosis and acceptance are inversely related. With respect to the breed trait, two of the
Table 2: Percentage of misclassifications and the number of selected features when polynomial kernel (SVMp ) and FSS methods are used. The three original data sets have 147
features.
RF
RF+FSS1
RF+FSS2
Error #Att. Error #Att. Error #Att.
tenderness
flavor
acceptance
20.0%
25.0%
24.7%
50.0
65.0
39.5
21.8%
26.5%
24.8%
27.0
33.5
30.0
21.3%
26.1%
25.3%
37.5
29.0
26.7
Avg.
23.2%
51.5
24.4%
30.2
24.2%
31.1
seven possible values, retinta and asturiana breeds [8], have more influence than others in
the preference function that describes consumer acceptance; for example, meat from retinta
animals seems to be the most appreciated by consumers.
In tenderness data set, the most useful attributes are: aging, fibrosis, residue and odor
intensity. Aging and fibrosis appear again, showing the relationship between acceptance
and tenderness. Residue depends on fibrosis, so it is not a surprise to find it in the list.
Apparently, odor intensity is not so related to tenderness, but it is closely related to aging.
5
Conclusions
We have shown that an approach based on nonlinear SVMs can be useful to model consumer preferences about beef meat. The polynomial model obtained and the FSS tools used
allow us to emphasize the relevance of meat traits previously described in the literature of
the field. However, the novelty of our approach is that we can algorithmically deduce the
expressions of relevance.
The sensory data base available probably tries to cover too many aspects of beef meat
affecting to its sensory quality. Therefore, it is not possible to obtain more detailed conclusions from the polynomial model. Nevertheless, the experience reported in this paper can
be very useful for the design of future experiments involving specific traits of beef meat
quality.
Acknowledgements
The research reported in this paper has been supported in part under Spanish Ministerio de
Ciencia y Tecnolog??a (MCyT) and Feder grant TIC2001-3579.
References
[1] A. Bahamonde, G. F. Bay?on, J. D??ez, J. R. Quevedo, O. Luaces, J. J. del Coz, J. Alonso, and
F. Goyache. Feature subset selection for learning preferences: A case study. In Proceedings
of the International Conference on Machine Learning, Banff, Alberta (Canada), July 2004.
Morgan Kaufmann.
[2] D. Buck, I. Wakeling, K. Greenhoff, and A. Hasted. Predicting paired preferences from sensory
data. Food quality and preference, 12:481?487, 2001.
[3] D. Corney. Designing food with bayesian belief networks. In Proceedings of the International
Conference on Adaptive Computing in engineering Design and Manufacture, pages 83?94,
2002.
[4] S. Degroeve, B. De Baets, Y. Van de Peer, and P. Rouz?e. Feature subset selection for splice site
prediction. Bioinformatics, 18(2):75?83, 2002.
[5] J. D??ez, A. Bahamonde, J. Alonso, S. L?opez, J. del Coz, J. Quevedo, J. Ranilla, O. Luaces,
?
I. Alvarez,
L. Royo, and F. Goyache. Artificial intelligence techniques point out differences
in classification performance between light and standard bovine carcasses. Meat Science,
64(3):249?258, 2003.
[6] J. D??ez, G. F. Bay?on, J. R. Quevedo, J. J. del Coz, O. Luaces, J. Alonso, and A. Bahamonde.
Discovering relevancies in very difficult regression problems: applications to sensory data analysis. In Proceedings of the European Conference on Artificial Intelligence (ECAI ?04), Valencia, Spain, 2004.
[7] S. Dumais, K. Bharat, T. Joachims, and A. Weigend, editors. Workshop on implicit measures
of user interests and preferences. In ACM SIGIR Conference, Toronto, Canada, 2003.
[8] M. Gil, X. Serra, M. Gispert, M. Oliver, C. Sa?nudo, B. Panea, J. Olleta, M. Campo, M. Oliv?an,
K. Osoro, M. Garcia-Cachan, R. Cruz-Sagredo, M. Izquierdo, M. Espejo, M. Mart??n, and
J. Piedrafita. The effect of breed-production systems on the myosin heavy chain 1, the biochemical characteristics and the colour variables of longissimus thoracis from seven spanish
beef cattle breeds. Meat Science, 58:181?188, 2001.
[9] F. Goyache, A. Bahamonde, J. Alonso, S. L?opez, del Coz J.J., J. Quevedo, J. Ranilla, O. Luaces,
I. Alvarez, L. Royo, and J. D??ez. The usefulness of artificial intelligence techniques to assess
subjective quality of products in the food industry. Trends in Food Science and Technology,
12(10):370?381, 2001.
[10] I. Guyon, J. Weston, S. Barnhill, and V. Vapnik. Gene selection for cancer classification using
support vector machines. Machine Learning, 46(1?3):389?422, 2002.
[11] R. Herbrich, T. Graepel, and K. Obermayer. Large margin rank boundaries for ordinal regression. In A. Smola, P. Bartlett, B. Scholkopf, and D. Schuurmans, editors, Advances in Large
Margin Classifiers, pages 115?132. MIT Press, Cambridge, MA, 2000.
[12] L. Jeremiah and L. Gibson. The effects of postmortem product handling and aging time on beef
palatability. Food Research International, 36:929?941, 2003.
[13] T. Joachims. Making large-scale support vector machines learning practical. In A. S.
B. Sch?olkopf, C. Burges, editor, Advances in Kernel Methods: Support Vector Machines. MIT
Press, Cambridge, MA, 1998.
[14] T. Joachims. Optimizing search engines using clickthrough data. In Proceedings of the ACM
Conference on Knowledge Discovery and Data Mining (KDD), 2002.
[15] J. Murray, C. Delahunty, and I. Baxter. Descriptive sensory analysis: past, present and future.
Food Research International, 36:461?471, 2001.
[16] T. N?s and E. Risvik. Multivariate analysis of data in sensory science. Elsevier, 1996.
[17] A. Rakotomamonjy. Variable selection using SVM-based criteria. Journal of Machine Learning
Research, 3:1357?1370, 2003.
[18] C. Sa?nudo, E. S. Macie, J. L. Olleta, M. Villaroel, B. Panea, and P. Albert??. The effects of
slaughter weight, breed type and ageing time on beef meat quality using two different texture
devices. Meat Science, 66:925?932, 2004.
[19] D. Schuurmans and F. Southey. Metric-based methods for adaptive model selection and regularization. Machine Learning, 48:51?84, 2002.
| 2720 |@word version:1 polynomial:8 seems:1 relevancy:1 reduction:2 score:2 att:3 subjective:2 past:1 comparing:1 adj:6 si:3 must:5 cruz:1 ministerio:1 kdd:1 remove:3 reproducible:1 designed:1 v:1 intelligence:4 selected:4 discovering:1 device:1 xk:3 cubist:3 toronto:1 preference:39 herbrich:2 banff:1 five:1 constructed:1 scholkopf:1 bharat:1 expected:3 behavior:1 frequently:1 alberta:1 food:9 considering:2 increasing:1 provided:1 discover:4 spain:1 panel:7 maximizes:2 what:1 kind:6 minimizes:2 developed:1 impractical:1 preferable:2 ensured:1 classifier:2 grant:1 appear:1 before:1 engineering:1 modify:1 aging:12 aiming:1 collect:1 someone:1 specialty:1 limited:1 discriminatory:1 practical:2 testing:5 recursive:1 x3:3 manufacture:1 gibson:1 induce:2 unlabeled:3 selection:15 context:1 influence:3 applying:1 restriction:1 map:1 center:1 straightforward:4 survey:1 sigir:1 simplicity:1 insight:2 handle:1 variation:1 laplace:1 target:1 commercial:1 infrequent:1 user:1 exact:1 designing:1 hypothesis:7 baets:1 trend:1 approximated:1 postmortem:1 role:1 solved:1 capture:1 ordering:5 decrease:2 highest:1 removed:1 environment:1 complexity:1 asked:2 productive:1 coz:5 ciencia:1 trained:1 uniformity:1 predictive:1 efficiency:1 f2:1 learner:1 acceptability:1 easily:1 represented:2 describe:1 artificial:4 peer:1 whose:3 kai:1 breed:14 final:2 sequence:6 advantage:1 descriptive:1 product:19 relevant:3 description:8 olkopf:1 rfe:4 assessing:1 r1:1 object:10 depending:1 measured:1 sa:2 quotient:1 come:1 concentrate:1 closely:2 correct:1 attribute:5 filter:3 additionaly:1 peculiarity:1 opinion:2 elimination:1 require:1 feeding:1 hx:3 f1:1 generalization:2 adjusted:2 correction:1 considered:1 predict:1 purpose:1 estimation:3 faithfully:1 create:2 tool:5 mit:2 clearly:1 sensor:1 fulfill:1 avoid:2 focus:1 joachim:3 improvement:1 rank:1 impossibility:1 industrial:1 summarizing:1 sense:1 elsevier:1 biochemical:1 reproduce:1 selects:1 bovine:4 overall:3 issue:1 classification:5 arg:1 among:2 animal:6 art:1 special:1 equal:1 once:1 construct:1 having:1 field:2 fibrosis:8 x4:3 represents:1 look:2 future:2 others:1 simplify:1 employ:1 producer:2 mcyt:1 randomly:1 individual:1 cheaper:1 replaced:2 acceptance:12 fd:1 interest:1 highly:1 mining:1 yielding:1 light:3 bundle:1 chain:1 accurate:2 oliver:1 capable:1 partial:1 experience:3 unless:1 rulequest:1 instance:1 industry:3 column:1 boolean:1 corroborate:1 cover:1 cost:2 deviation:2 subset:14 rakotomamonjy:1 predictor:2 usefulness:1 too:1 reported:5 considerably:1 dumais:1 chunk:1 international:4 participates:1 together:1 again:2 containing:1 choose:1 worse:1 pretended:1 expert:9 derivative:1 return:4 potential:1 de:4 summarized:1 includes:1 explicitly:1 ranking:13 depends:1 piece:5 performed:1 try:3 view:2 apparently:1 carlos:1 capability:1 ass:2 formed:2 accuracy:4 kaufmann:1 variance:1 who:1 characteristic:3 judgment:11 identify:2 repeatability:1 yield:1 myosin:1 raw:1 bayesian:1 explain:2 opez:2 barnhill:1 whenever:1 definition:1 bahamonde:5 acquisition:2 couple:1 gain:2 sampled:1 duplicated:1 knowledge:3 organized:2 graepel:1 sophisticated:1 actually:1 higher:8 dt:2 day:2 follow:1 specify:1 alvarez:2 just:2 implicit:2 smola:1 until:1 d:2 hand:4 working:1 ei:1 replacing:1 nonlinear:1 assessment:1 del:5 quality:9 aj:3 effect:5 true:1 regularization:1 chemical:2 iteratively:1 zaragoza:1 deal:2 spanish:5 cosine:1 criterion:2 trying:1 impression:2 svmp:5 image:1 meaning:1 fi:4 common:1 labour:1 physical:3 trait:12 interpret:2 kwk2:1 cambridge:2 cv:1 ai:2 rd:4 fk:3 session:8 pointed:1 had:1 similarity:3 etc:1 base:2 deduce:1 multivariate:1 own:1 perspective:1 optimizing:1 scenario:1 certain:1 binary:1 success:1 yi:4 morgan:1 campo:1 novelty:1 period:3 july:1 adapt:1 cross:2 paired:1 prediction:4 involving:1 regression:17 metric:3 albert:1 iteration:1 kernel:13 sometimes:1 represent:1 achieved:2 addition:2 residue:2 affecting:1 crucial:1 sch:1 rest:1 specially:1 probably:1 induced:1 valencia:1 near:3 noting:1 svmlight:1 baxter:1 xj:3 zi:1 misclassifications:4 opposite:1 inner:1 idea:3 reduce:3 consumed:1 expression:1 bartlett:1 feder:1 colour:1 forecasting:1 f:4 adequate:1 useful:8 buck:1 clear:1 detailed:1 amount:1 prepared:1 processed:1 category:1 svms:1 percentage:2 zj:2 notice:3 gil:1 estimated:1 svml:2 algorithmically:1 express:3 group:7 key:1 redundancy:3 nevertheless:1 weigend:1 almost:1 guyon:1 coherence:1 cachan:1 scaling:2 kaj:1 cattle:1 fl:5 def:2 played:1 aic:1 fold:1 adapted:1 ri:3 x2:3 sake:1 aspect:2 speed:1 min:1 according:1 across:1 slightly:1 describes:1 making:2 equation:2 previously:5 describing:2 know:1 ordinal:1 end:1 available:2 operation:1 endowed:1 observe:1 disagreement:4 batch:3 odor:3 original:6 denotes:1 include:1 calculating:1 uj:3 establish:1 build:1 murray:1 objective:2 already:1 slaughter:2 traditional:1 hai:1 obermayer:1 gradient:1 distance:7 unable:1 capacity:1 alonso:4 seven:2 reason:3 ageing:1 consumer:34 carcass:5 relationship:2 difficult:3 expense:1 rise:1 design:2 clickthrough:1 perform:1 appreciation:1 allowing:1 defining:1 precise:1 tenderness:11 worthy:1 intensity:2 canada:2 rating:17 introduced:3 pair:7 engine:1 learned:1 able:1 usually:5 rf:9 max:1 belief:1 suitable:2 satisfaction:3 predicting:1 improve:4 rated:1 technology:1 inversely:1 literature:1 taste:1 acknowledgement:1 discovery:1 relative:3 loss:1 interesting:1 validation:2 southey:1 degree:4 editor:3 heavy:1 production:1 cancer:1 supported:1 copy:1 ecai:1 appreciated:2 bias:1 formal:1 allow:2 burges:1 absolute:4 serra:1 distributed:1 van:1 boundary:1 calculated:1 dimension:1 sensory:22 made:2 avg:2 adaptive:2 selector:1 meat:30 emphasize:1 gene:1 dealing:2 conceptual:1 xi:3 search:1 iterative:2 bay:3 table:5 additionally:3 learn:4 zk:2 correlated:1 schuurmans:2 improving:1 quevedo:4 beef:17 du:1 untrained:2 excellent:1 necessarily:1 separator:1 constructing:1 european:1 main:3 repeated:1 fair:1 x1:3 site:1 espejo:1 splice:1 formula:1 removing:1 down:1 specific:1 showing:1 list:3 r2:2 svm:8 experimented:1 essential:1 intrinsic:1 workshop:1 vapnik:1 importance:1 texture:2 margin:2 flavor:4 surprise:1 garcia:1 simply:1 ez:5 expressed:1 ordered:2 partially:1 nested:2 relies:1 acm:2 mart:1 ma:2 weston:1 viewed:1 goal:1 uniformly:1 called:1 gij:1 e:2 experimental:2 indicating:1 select:2 people:2 support:4 assessed:3 meant:1 relevance:4 bioinformatics:1 requisite:1 tested:7 handling:1 |
1,895 | 2,721 | Multiple Alignment of Continuous Time Series
Jennifer Listgarten? , Radford M. Neal? , Sam T. Roweis? and Andrew Emili?
Department of Computer Science, ? Banting and Best Department of Medical Research
and Program in Proteomics and Bioinformatics
University of Toronto, Toronto, Ontario, M5S 3G4
{jenn,radford,roweis}@cs.toronto.edu, [email protected]
?
Abstract
Multiple realizations of continuous-valued time series from a stochastic
process often contain systematic variations in rate and amplitude. To
leverage the information contained in such noisy replicate sets, we need
to align them in an appropriate way (for example, to allow the data to be
properly combined by adaptive averaging). We present the Continuous
Profile Model (CPM), a generative model in which each observed time
series is a non-uniformly subsampled version of a single latent trace, to
which local rescaling and additive noise are applied. After unsupervised
training, the learned trace represents a canonical, high resolution fusion
of all the replicates. As well, an alignment in time and scale of each
observation to this trace can be found by inference in the model. We
apply CPM to successfully align speech signals from multiple speakers
and sets of Liquid Chromatography-Mass Spectrometry proteomic data.
1
A Profile Model for Continuous Data
When observing multiple time series generated by a noisy, stochastic process, large systematic sources of variability are often present. For example, within a set of nominally
replicate time series, the time axes can be variously shifted, compressed and expanded,
in complex, non-linear ways. Additionally, in some circumstances, the scale of the measured data can vary systematically from one replicate to the next, and even within a given
replicate.
We propose a Continuous Profile Model (CPM) for simultaneously analyzing a set of such
time series. In this model, each time series is generated as a noisy transformation of a
single latent trace. The latent trace is an underlying, noiseless representation of the set
of replicated, observable time series. Output time series are generated from this model
by moving through a sequence of hidden states in a Markovian manner and emitting an
observable value at each step, as in an HMM. Each hidden state corresponds to a particular
location in the latent trace, and the emitted value from the state depends on the value of
the latent trace at that position. To account for changes in the amplitude of the signals
across and within replicates, the latent time states are augmented by a set of scale states,
which control how the emission signal will be scaled relative to the value of the latent trace.
During training, the latent trace is learned, as well as the transition probabilities controlling
the Markovian evolution of the scale and time states and the overall noise level of the
observed data. After training, the latent trace learned by the model represents a higher
resolution ?fusion? of the experimental replicates. Figure 1 illustrate the model in action.
Unaligned, Linear Warp Alignment and CPM Alignment
Amplitude
40
30
20
10
0
50
Amplitude
40
30
20
10
Amplitude
0
30
20
10
0
Time
a)
b)
Figure 1: a) Top: ten replicated speech energy signals as described in Section 4), Middle:
same signals, aligned using a linear warp with an offset, Bottom: aligned with CPM (the
learned latent trace is also shown in cyan). b) Speech waveforms corresponding to energy
signals in a), Top: unaligned originals, Bottom: aligned using CPM.
2
Defining the Continuous Profile Model (CPM)
The CPM is generative model for a set of K time series, ~xk = (xk1 , xk2 , ..., xkN k ). The
temporal sampling rate within each ~xk need not be uniform, nor must it be the same across
the different ~xk . Constraints on the variability of the sampling rate are discussed at the end
of this section. For notational convenience, we henceforth assume N k = N for all k, but
this is not a requirement of the model.
The CPM is set up as follows: We assume that there is a latent trace, ~z = (z1 , z2 , ..., zM ), a
canonical representation of the set of noisy input replicate time series. Any given observed
time series in the set is modeled as a non-uniformly subsampled version of the latent trace
to which local scale transformations have been applied. Ideally, M would be infinite, or at
least very large relative to N so that any experimental data could be mapped precisely to the
correct underlying trace point. Aside from the computational impracticalities this would
pose, great care to avoid overfitting would have to be taken. Thus in practice, we have used
M = (2 + )N (double the resolution, plus some slack on each end) in our experiments
and found this to be sufficient with < 0.2. Because the resolution of the latent trace is
higher than that of the observed time series, experimental time can be made effectively to
speed up or slow down by advancing along the latent trace in larger or smaller jumps.
The subsampling and local scaling used during the generation of each observed time series are determined by a sequence of hidden state variables. Let the state sequence for
observation k be ~? k . Each state in the state sequence maps to a time state/scale state pair:
?ik ? {?ik , ?ki }. Time states belong to the integer set (1..M ); scale states belong to an
ordered set (?1 ..?Q ). (In our experiments we have used Q=7, evenly spaced scales in
logarithmic space). States, ?ik , and observation values, xki , are related by the emission
probability distribution: A?ik (xki |~z) ? p(xki |?ik , ~z, ?, uk ) ? N (xki ; z?ik ?ki uk , ?), where ?
is the noise level of the observed data, N (a; b, c) denotes a Gaussian probability density
for a with mean b and standard deviation c. The uk are real-valued scale parameters, one
per observed time series, that correct for any overall scale difference between time series k
and the latent trace.
To fully specify our model we also need to define the state transition probabilities. We
define the transitions between time states and between scale states separately, so that
T?ki?1 ,?i ? p(?i |?i?1 ) = p(?i |?i?1 )pk (?i |?i?1 ). The constraint that time must move
forward, cannot stand still, and that it can jump ahead no more than J? time states is enforced. (In our experiments we used J? = 3.) As well, we only allow scale state transitions
between neighbouring scale states so that the local scale cannot jump arbitrarily. These
constraints keep the number of legal transitions to a tractable computational size and work
well in practice. Each observed time series has its own time transition probability distribution to account for experiment-specific patterns. Both the time and scale transition
probability distributions are given by multinomials:
?
?
dk1 , if a ? b = 1
?
?
?dk , if a ? b = 2
?
?
? 2
k
p (?i = a|?i?1 = b) = ...
?
?
?
?dkJ? , if a ? b = J?
?
?
?0,
otherwise
p(?i = a|?i?1
?
s0 , if D(a, b) = 0
?
?
?
s1 , if D(a, b) = 1
= b) =
?s1 , if D(a, b) = ?1
?
?
0, otherwise
where D(a, b) = 1 means that a is one scale state larger than b, and D(a, b) = ?1 means
that a is one scale state smaller than b, and D(a, b) = 0 means that a = b. The distributions
PJ ? k
are constrained by: i=1
di = 1 and 2s1 + s0 = 1.
J? determines the maximum allowable instantaneous speedup of one portion of a time
series relative to another portion, within the same series or across different series. However,
the length of time for which any series can move so rapidly is constrained by the length of
the latent trace; thus the maximum overall ratio in speeds achievable by the model between
any two entire time series is given by min(J? , M
N ).
After training, one may examine either the latent trace or the alignment of each observable
time series to the latent trace. Such alignments can be achieved by several methods, including use of the Viterbi algorithm to find the highest likelihood path through the hidden
states [1], or sampling from the posterior over hidden state sequences. We found Viterbi
alignments to work well in the experiments below; samples from the posterior looked quite
similar.
3
Training with the Expectation-Maximization (EM) Algorithm
As with HMMs, training with the EM algorithm (often referred to as Baum-Welch in the
context of HMMs [1]), is a natural choice. In our model the E-Step is computed exactly
using the Forward-Backward algorithm [1], which provides the posterior probability over
states for each time point of every observed time series, ?sk (i) ? p(?i = s|~x) and also
the pairwise state posteriors, ?s,t (i) ? p(?i?1 = s, ?i = t|~xk ). The algorithm is modified
only in that the emission probabilities depend on the latent trace as described in Section 2.
The M-Step consists of a series of analytical updates to the various parameters as detailed
below.
Given the latent trace (and the emission and state transition probabilities), the complete log
likelihood of K observed time series, ~xk , is given by Lp ? L + P. L is the likelihood term
arising in a (conditional) HMM model, and can be obtained from the Forward-Backward
algorithm. It is composed of the emission and state transition terms. P is the log prior
(or penalty term), regularizing various aspects of the model parameters as explained below.
These two terms are:
!
K
N
N
X
X
X
k
k
L?
log T?i?1 ,?i
log p(?1 ) +
log A?i (xi |~z) +
(1)
P ? ??
i=2
i=1
k=1
?
?1
X
(zj+1 ? zj )2 +
j=1
K
X
log D(dkv |{?vk }) + log D(sv |{?v0 }),
(2)
k=1
where p(?1 ) are priors over the initial states. The first term in Equation 2 is a smoothing
penalty on the latent trace, with ? controlling the amount of smoothing. ?vk and ?v0 are
Dirichlet hyperprior parameters for the time and scale state transition probability distributions respectively. These ensure that all non-zero transition probabilities remain non-zero.
For the time state transitions, v ? {1, J? } and ?vk corresponds to the pseudo-count data for
the parameters d1 , d2 . . . dJ? . For the scale state transitions, v ? {0, 1} and ?vk corresponds
to the pseudo-count data for the parameters s0 and s1 .
Letting S be the total number of possible states, that is, the number of elements in the
cross-product of possible time states and possible scale states, the expected complete log
likelihood is:
p
<L >? =P +
K X
S
X
k
?sk (1) log T0,s
+
k=1 s=1
... +
K X
S
X
K X
S X
N
X
?sk (i) log As (xki |~z) + . . .
k=1 s=1 i=1
S X
N
X
k
k
?s,s
0 (i) log Ts,s0
k=1 s=1 s0 =1 i=2
k
k
using the notation T0,s
? p(?1 = s), and where ?sk (i) and ?s,s
0 (i) are the posteriors over
states as defined above. Taking derivatives of this quantity with respect to each of the
parameters and finding the critical points provides us with the M-Step update equations. In
updating the latent trace ~z we obtain a system of M simultaneous equations, for j = 1..M :
K
X
? <Lp>?
=0=
?zj
X
N
X
?sk (i)?s uk
k=1 {s|?s =j} i=1
(xki ? zj uk ?s )
? ?(4zj ? 2zj?1 ? 2zj+1 )
?2
For the cases j = 1, N , the terms 2zj?1 and 2zj+1 , respectively, drop out. Considering
all such equations we obtain a system of M equations in M unknowns. Each equation
depends only linearly on three variables from the latent trace. Thus the solution is easily
obtained numerically by solving a tridiagonal linear system.
Analytic updates for ? 2 and uk are given by:
2
? =
PS
s=1
PN
i=1
?sk (i)(xki ? z?s uk ?s )2
,
N
k
u =
PS
PN k
k
s=1 z?s ?s
i=1 ?s (i)xi
PS
P
N
2
k
s=1 (z?s ?s )
i=1 ?s (i)
Lastly, updates for the scale and state transition probability distributions are given by:
dkv
= PJ ?
j=1
s v = P1
j=0
?vk +
?jk +
?j0 +
?j0 +
PS
s=1
PJ ?
j=1
PN k
{s0 |?s0 ??s =v}
i=2 ?s,s00 (i)
PS P
PN k
s=1
{s0 |?s0 ??s =j}
i=2 ?s,s00 (i)
P
PK P S
k=1
PK
k=1
s=1
PS
s=1
PN
k
i=2 ?s,s00 (i)
P
PN k
{s00 ?H(s,1),H(s,0)}
i=2 ?s,s00 (i)
P
{s00 ?H(s,v)}
where H(s, j) ? {s0 |s0 is exactly j scale states away from s}.
Note that we do
not normalize the Dirichlets, and omit the traditional minus one in the exponent:
QJ?
Q1
0
k
D(dkv |{?vk }) = v=1
(dkv )?v and D(sv |{?v0 }) = v=0 (sv )?v .
The M-Step updates uk , ?, and ~z are coupled. Thus we arbitrarily pick an order to update
them and as one is updated, its new values are used in the updates for the coupled parameter
updates that follow it. In our experiments we updated in the following order: ?, ~z, u k . The
other two parameters, dkv and sv , are completely decoupled.
4
Experiments with Laboratory and Speech Data
We have applied the CPM model to analyze several Liquid Chromatography - Mass Spectrometry (LC-MS) data sets from an experimental biology laboratory. Mass spectrometry
technology is currently being developed to advance the field of proteomics [2, 3]. A mass
spectrometer takes a sample as input, for example, human blood serum, and produces a
measure of the abundance of molecules that have particular mass/charge ratios. In proteomics the molecules in question are small protein fragments. From the pattern of abundance values one can hope to infer which proteins are present and in what quantity. For
protein mixtures that are very complex, such as blood serum, a sample preparation step
is used to physically separate parts of the sample on the basis of some property of the
molecules, for example, hydrophobicity. This separation spreads out the parts over time
so that at each unique time point a less complex mixture is fed into the mass spectrometer.
The result is a two-dimensional time series spectrum with mass/charge on one axis and
time of input to the mass spectrometer on the other. In our experiments we collapsed the
data at each time point to one dimension by summing together abundance values over all
mass/charge values. This one-dimensional data is referred to as the Total Ion Count (TIC).
We discuss alternatives to this in the last section. After alignment of the TICs, we assessed
the alignment of the LC-MS data by looking at both the TIC alignments, and also the corresponding two-dimensional alignments of the non-collapsed data, which is where the true
information lies.
The first data set was a set of 13 replicates, each using protein extracted from lysed E. coli
cells. Proteins were digested and subjected to capillary-scale LC-MS coupled on-line to an
ion trap mass spectrometer. First we trained the model with no smoothing (i.e., ? = 0) on
the 13 replicates. This provided nice alignments when viewed in both the TIC space and
the full two-dimensional space. Next we used leave-one-out cross-validation on six of the
replicates in order to choose a suitable value for ?. Because the uk and dkv are time series
specific, we ran a restricted EM on the hold-out case to learn these parameters, holding the
other parameters fixed at the values found from learning on the training set. Sixteen values
of ? over five orders of magnitude, and also zero, were used. Note that we did not include
the regularization likelihood term in the calculations of hold-out likelihood. One of the
non-zero values was found to be optimal (statistically significant at a p=0.05 level using a
paired sample t-test to compare it to no smoothing). Visually, there did not appear to be
a difference between no regularization and the optimal value of ?, in either the TIC space
or the full two-dimensional space. Figure 2 shows the alignments applied to the TICs and
also the two-dimensional data, using the optimal value of ?.
Unaligned and Aligned Time Series
8
x 10
Replicate 5
8
x 10
9
Latent Trace
Aligned Experimental Time Series
Original Time Series
8
8
7
6
6
Amplitude
Amplitude
10
4
2
5
4
3
8
0 x 10
Latent Space Amplitude
2
1
6
0
100
Residual
200
300
400
500
600
700
800
4
3
2
2
Time Jump From Previous State
1
Scale States
0
100
200
300
400
500
Time
600
700
800
200
400
Latent Time
a)
b)
c)
d)
600
800
Figure 2: Figure 2: a) Top: 13 Replicate pre-processed TICs as described in Section 4),
Bottom: same as top, but aligned with CPM (the learned latent trace is also shown). b) The
fifth TIC replicate aligned to the learned latent trace (inset shows the original, unaligned).
Below are three strips showing, from top-to-bottom, i) the error residual, ii) the number
of time states moved between every two states in the Viterbi alignment, and iii) the local
scaling applied at each point in the alignment. c) A portion of the two-dimensional LC-MS
data from replicates two (in red) and four (in green). d) Same as c), but after alignment (the
same one dimensional alignment was applied to every Mass/Charge value). Marker lines
labeled A to F show how time in c) was mapped to latent time using the Viterbi alignment.
We also trained our model on five different sets of LC-MS data, each consisting of human
blood serum. We used no smoothing and found the results visually similar in quality to the
first data set.
To ensure convergence to a good local optimum and to speed up training, we pre-processed
the LC-MS data set by coarsely aligning and scaling each time series as follows: We 1)
translated each time series so that the center of mass of each time series was aligned to the
median center of mass over all time series, 2) scaled the abundance values such that the
sum of abundance values in each time series was equal to the median sum of abundance
values over all time series.
We also used our model to align 10 speech signals, each an utterance of the same sentence
spoken by a different speaker. The short-time energy (using a 30ms Hanning window) was
computed every 8ms for each utterance and the resulting vectors were used as the input
to CPM for alignment. The smoothing parameter ? was set to zero. For comparison, we
also performed a linear warping of time with an offset. (i.e. each signal was translated so
as to start at the same time, and the length of each signal was stretched or compressed so
as to each occupy the same amount of time). Figure 1 shows the successful alignment of
the speech signals by CPM and also the (unsuccessful) linear warp. Audio for this example can be heard at http://www.cs.toronto.edu/?jenn/alignmentStudy,
which also contains some supplemental figures for the paper.
Initialization for EM training was performed as follows: ? was set to 15% of the difference
between the maximum and minimum values of the first time series. The latent trace was
initialized to be the first observed time series, with Gaussian, zero-mean noise added, with
standard deviation equal to ?. This was then upsampled by a factor of two by repeating
every value twice in a row. The additional slack at either end of the latent trace was set to
be the minimum value seen in the given time series. The uk were each set to one and the
multinomial scale and state transition probabilities were set to be uniform.
5
Related Algorithms and Models
Our proposed CPM has many similarities to Input/Output HMMs (IOHMMs), also called
Conditional HMMs [4]. IOHMMs extend standard HMMs [1] by conditioning the emission
and transition probabilities on an observed input sequence. Each component of the output
sequence corresponds to a particular component of the input. Training of an IOHMM
is supervised ? a mapping from an observed input sequence to output target sequence
is learned. Our CPM also requires input and thus is also a type of conditional HMM.
However, the input is unobserved (but crucially it is shared between all replicates) and
hence learning is unsupervised in the CPM model. One could also take the alternative view
that the CPM is simply an HMM with an extra set of parameters, the latent trace, that affect
the emission probabilities and which are learned by the model.
The CPM is similar in spirit to Profile HMMs which have been used with great success
for discrete, multiple sequence alignment, modeling of protein families and their conserved structures, gene finding [5], among others. Profile HMM are HMMs augmented
by constrained-transition ?Delete? and ?Insert? states, with the former emitting no observations. Multiple sequences are provided to the Profile HMM during training and a summary
of their shared statistical properties is contained in the resulting model. The development
of Profile HMMs has provided a robust, statistical framework for reasoning about sets of
related discrete sequence data. We put forth the CPM as a continuous data, conditional
analogue.
Many algorithms currently used for aligning continuous time series data are variations of
Dynamic Time Warping (DTW) [6], a dynamic programming based approach which originated in the speech recognition community as a robust distance measure between two time
series. DTW works on pairs of time series, aligning one time series to a specified reference
time series. DTW does not take in to account systematic variations in the amplitude of the
signal. Our CPM can be viewed as a rich and robust extension of DTW that can be applied
to many time series in parallel and which automatically uncovers the underlying template
of the data.
6
Discussion and Conclusion
We have introduced a generative model for sets of continuous, time series data. By training
this model one can leverage information contained in noisy, replicated experimental data,
and obtain a single, superior resolution ?fusion? of the data. We demonstrated successful
use of this model on real data, but note that it could be applied to a wide range of problems
involving time signals, for example, alignment of gene expression time profiles, alignment
of temporal physiological signals, alignment of motion capture data, to name but a few.
Certain assumptions of the model presented here may be violated under different experimental conditions. For example, the Gaussian emission probabilities treat errors in
large amplitudes in the same absolute terms as in smaller amplitudes, whereas in reality, it may be that the error scales with signal amplitude. Similarly, the penalty term
P? ?1
?? j=1 (zj+1 ? zj )2 does not scale with the amplitude; this might result in the model
arbitrarily preferring a lower amplitude latent trace. (However, in practice, we did not find
this to be a problem.)
One immediate and straight-forward extension to the model would be to allow the data at
each time point to be a multi-dimensional feature vector rather than a scalar value. This
could easily be realized by allowing the emission probabilities to be multi-dimensional. In
this way a richer set of information could be used: either the raw, multi-dimensional feature
vector, or some transformation of the feature vectors, for example, Principal Components
Analysis. The rest of the model would be unchanged and each feature vector would move
as a coherent piece. However, it might also be useful to allow different dimensions of the
feature vector to be aligned differently. For example, with the LC-MS data, this might
mean allowing different mass/charge peptides to be aligned differently at each time point.
However, in its full generality, such a task would be extremely computational intense.
A perhaps more interesting extension is to allow the model to work with non-replicate data.
For example, suppose one had a set of LC-MS experiments from a set of cancer patients,
and also a set from normal persons. It would be desirable to align the whole set of time
series and also to have the model tease out the differences between them. One approach
is to consider the model to be semi-supervised - the model is told the class membership
of each training example. Then each class is assigned its own latent trace, and a penalty
is introduced for any disagreements between the latent traces. Care needs to be taken to
ensure that the penalty plateaus after a certain amount of disagreement between latent trace
points, so that parts of the latent trace which are truly different are able to whole-heartedly
disagree. Assuming that the time resolution in the observed time series is sufficiently high,
one might also want to encourage the amount of disagreement over time to be Markovian.
That is, if the previous time point disagreed with the other latent traces, then the current
point should be more likely to disagree.
References
[1] Alan B. Poritz. Hidden markov models: A guided tour. In Proceedings of the IEEE Conference
on Acoustics, Speech and Signal Processing (ICASSP), pages 7?13. Morgan Kaufmann, 1988.
[2] Ruedi Aebersold and Matthias Mann. Mass spectrometry-based proteomics. Nature, 422:198?
207, 2003.
[3] P. Kearney and P. Thibault. Bioinformatics meets proteomics - bridging the gap between mass
spectrometry data analysis and cell biology. Journal of Bioinformatics and Computational Biology, 1:183?200, 2003.
[4] Yoshua Bengio and Paolo Frasconi. An input output HMM architecture. In G. Tesauro, D. Touretzky, and T. Leen, editors, Advances in Neural Information Processing Systems, volume 7, pages
427?434. The MIT Press, 1995.
[5] Richard Durbin, Sean R. Eddy, Anders Krogh, and Graeme Mitchison. Biological Sequence
Analysis: Probabilistic Models of Proteins and Nucleic Acids. Cambridge Univ. Press, 2000.
Durbin.
[6] H. Sakoe and S.Chiba. Dynamic programming algorithm for spoken word recognition. Readings
in Speech Recognition, pages 159?165, 1990.
| 2721 |@word middle:1 version:2 achievable:1 replicate:9 d2:1 crucially:1 uncovers:1 q1:1 pick:1 minus:1 xkn:1 initial:1 series:49 fragment:1 contains:1 liquid:2 current:1 z2:1 must:2 additive:1 analytic:1 drop:1 update:8 aside:1 poritz:1 generative:3 xk:5 short:1 provides:2 toronto:4 location:1 five:2 along:1 ik:6 consists:1 manner:1 sakoe:1 g4:1 pairwise:1 expected:1 p1:1 nor:1 examine:1 multi:3 automatically:1 ruedi:1 window:1 considering:1 provided:3 underlying:3 dk1:1 notation:1 mass:16 what:1 tic:8 developed:1 supplemental:1 spoken:2 finding:2 transformation:3 unobserved:1 temporal:2 pseudo:2 every:5 charge:5 exactly:2 scaled:2 uk:10 control:1 medical:1 omit:1 appear:1 local:6 treat:1 analyzing:1 meet:1 path:1 might:4 plus:1 twice:1 initialization:1 hmms:8 range:1 statistically:1 unique:1 practice:3 j0:2 pre:2 word:1 upsampled:1 protein:7 convenience:1 cannot:2 put:1 context:1 collapsed:2 www:1 map:1 demonstrated:1 center:2 baum:1 serum:3 resolution:6 impracticality:1 welch:1 variation:3 updated:2 controlling:2 target:1 suppose:1 neighbouring:1 programming:2 element:1 recognition:3 jk:1 updating:1 labeled:1 observed:14 bottom:4 capture:1 spectrometer:4 highest:1 ran:1 ideally:1 dynamic:3 trained:2 depend:1 solving:1 completely:1 basis:1 translated:2 easily:2 icassp:1 differently:2 various:2 univ:1 banting:1 quite:1 richer:1 larger:2 valued:2 otherwise:2 compressed:2 emili:2 noisy:5 listgarten:1 sequence:13 analytical:1 matthias:1 propose:1 unaligned:4 product:1 zm:1 aligned:10 realization:1 rapidly:1 ontario:1 roweis:2 aebersold:1 graeme:1 forth:1 moved:1 normalize:1 convergence:1 double:1 requirement:1 p:6 optimum:1 produce:1 leave:1 illustrate:1 andrew:2 pose:1 measured:1 krogh:1 c:2 waveform:1 proteomic:1 guided:1 correct:2 stochastic:2 human:2 mann:1 biological:1 insert:1 extension:3 hold:2 sufficiently:1 normal:1 visually:2 great:2 viterbi:4 mapping:1 vary:1 xk2:1 cpm:20 currently:2 peptide:1 successfully:1 hope:1 mit:1 gaussian:3 modified:1 rather:1 avoid:1 pn:6 ax:1 emission:9 properly:1 notational:1 vk:6 likelihood:6 inference:1 anders:1 membership:1 iohmm:1 entire:1 hidden:6 overall:3 among:1 exponent:1 development:1 constrained:3 smoothing:6 field:1 equal:2 frasconi:1 sampling:3 biology:3 represents:2 unsupervised:2 others:1 yoshua:1 richard:1 few:1 composed:1 simultaneously:1 variously:1 subsampled:2 consisting:1 alignment:24 replicates:8 mixture:2 truly:1 encourage:1 decoupled:1 intense:1 hyperprior:1 initialized:1 delete:1 modeling:1 markovian:3 maximization:1 deviation:2 tour:1 uniform:2 successful:2 tridiagonal:1 sv:4 combined:1 person:1 density:1 preferring:1 systematic:3 told:1 probabilistic:1 together:1 dirichlets:1 s00:6 choose:1 henceforth:1 coli:1 derivative:1 rescaling:1 account:3 depends:2 piece:1 performed:2 view:1 observing:1 analyze:1 portion:3 red:1 start:1 parallel:1 capillary:1 kaufmann:1 acid:1 spaced:1 raw:1 m5s:1 straight:1 simultaneous:1 plateau:1 touretzky:1 strip:1 energy:3 di:1 eddy:1 amplitude:14 sean:1 higher:2 supervised:2 follow:1 specify:1 leen:1 generality:1 xk1:1 lastly:1 marker:1 quality:1 perhaps:1 name:1 contain:1 dkj:1 true:1 evolution:1 former:1 regularization:2 hence:1 assigned:1 laboratory:2 neal:1 during:3 speaker:2 m:10 allowable:1 complete:2 motion:1 reasoning:1 regularizing:1 instantaneous:1 superior:1 multinomial:2 conditioning:1 volume:1 discussed:1 belong:2 extend:1 numerically:1 significant:1 cambridge:1 stretched:1 similarly:1 digested:1 dj:1 had:1 moving:1 similarity:1 v0:3 align:4 aligning:3 posterior:5 own:2 tesauro:1 certain:2 arbitrarily:3 success:1 conserved:1 seen:1 minimum:2 additional:1 care:2 morgan:1 signal:15 ii:1 semi:1 multiple:6 full:3 desirable:1 infer:1 thibault:1 alan:1 calculation:1 cross:2 paired:1 involving:1 proteomics:5 circumstance:1 noiseless:1 expectation:1 physically:1 patient:1 achieved:1 ion:2 cell:2 spectrometry:5 whereas:1 want:1 separately:1 median:2 source:1 extra:1 rest:1 spirit:1 emitted:1 integer:1 leverage:2 iii:1 bengio:1 affect:1 architecture:1 qj:1 t0:2 six:1 expression:1 bridging:1 penalty:5 speech:9 action:1 useful:1 heard:1 detailed:1 amount:4 repeating:1 ten:1 processed:2 http:1 occupy:1 canonical:2 zj:11 shifted:1 arising:1 per:1 discrete:2 coarsely:1 paolo:1 four:1 blood:3 pj:3 advancing:1 backward:2 sum:2 enforced:1 family:1 separation:1 scaling:3 cyan:1 ki:3 durbin:2 ahead:1 constraint:3 precisely:1 aspect:1 speed:3 xki:7 min:1 extremely:1 expanded:1 speedup:1 department:2 across:3 smaller:3 em:4 sam:1 remain:1 lp:2 s1:4 explained:1 restricted:1 taken:2 legal:1 equation:6 jennifer:1 slack:2 count:3 discus:1 letting:1 tractable:1 fed:1 end:3 subjected:1 apply:1 away:1 appropriate:1 disagreement:3 alternative:2 original:3 top:5 denotes:1 subsampling:1 dirichlet:1 ensure:3 include:1 unchanged:1 warping:2 move:3 question:1 quantity:2 looked:1 added:1 realized:1 traditional:1 distance:1 separate:1 mapped:2 hmm:7 evenly:1 assuming:1 length:3 modeled:1 ratio:2 holding:1 trace:37 unknown:1 allowing:2 disagree:2 observation:4 nucleic:1 markov:1 iohmms:2 t:1 immediate:1 defining:1 variability:2 looking:1 community:1 introduced:2 pair:2 specified:1 z1:1 sentence:1 acoustic:1 coherent:1 learned:8 able:1 below:4 pattern:2 reading:1 program:1 including:1 green:1 unsuccessful:1 analogue:1 critical:1 suitable:1 natural:1 residual:2 technology:1 dtw:4 axis:1 coupled:3 utterance:2 prior:2 nice:1 relative:3 fully:1 generation:1 interesting:1 sixteen:1 hydrophobicity:1 validation:1 sufficient:1 s0:11 editor:1 systematically:1 row:1 cancer:1 summary:1 last:1 tease:1 allow:5 warp:3 wide:1 template:1 taking:1 fifth:1 absolute:1 chiba:1 dimension:2 transition:17 stand:1 rich:1 forward:4 made:1 adaptive:1 replicated:3 jump:4 emitting:2 observable:3 keep:1 gene:2 overfitting:1 summing:1 xi:2 mitchison:1 spectrum:1 continuous:9 latent:38 sk:6 reality:1 additionally:1 learn:1 nature:1 molecule:3 ca:1 robust:3 complex:3 did:3 pk:3 spread:1 linearly:1 whole:2 noise:4 profile:9 augmented:2 chromatography:2 referred:2 slow:1 lc:8 position:1 hanning:1 originated:1 lie:1 abundance:6 down:1 specific:2 utoronto:1 inset:1 showing:1 offset:2 dk:1 physiological:1 fusion:3 trap:1 disagreed:1 effectively:1 jenn:2 magnitude:1 gap:1 logarithmic:1 simply:1 likely:1 contained:3 ordered:1 scalar:1 nominally:1 radford:2 corresponds:4 determines:1 extracted:1 conditional:4 viewed:2 shared:2 change:1 infinite:1 determined:1 uniformly:2 averaging:1 principal:1 total:2 called:1 experimental:7 assessed:1 bioinformatics:3 violated:1 preparation:1 audio:1 d1:1 |
1,896 | 2,722 | Learning, Regularization and Ill-Posed Inverse
Problems
Lorenzo Rosasco
DISI, Universit`a di Genova
Genova, I
[email protected]
Ernesto De Vito
Dipartimento di Matematica
Universit`a di Modena
and INFN, Sezione di Genova
Genova, I
[email protected]
Andrea Caponnetto
DISI, Universit`a di Genova
Genova, I
[email protected]
Umberto De Giovannini
DISI, Universit`a di Genova
Genova, I
[email protected]
Francesca Odone
DISI, Universit`a di Genova
Genova, I
[email protected]
Abstract
Many works have shown that strong connections relate learning from examples to regularization techniques for ill-posed inverse problems. Nevertheless by now there was no formal evidence neither that learning from
examples could be seen as an inverse problem nor that theoretical results
in learning theory could be independently derived using tools from regularization theory. In this paper we provide a positive answer to both
questions. Indeed, considering the square loss, we translate the learning
problem in the language of regularization theory and show that consistency results and optimal regularization parameter choice can be derived
by the discretization of the corresponding inverse problem.
1
Introduction
The main goal of learning from examples is to infer an estimator, given a finite sample
of data drawn according to a fixed but unknown probabilistic input-output relation. The
desired property of the selected estimator is to perform well on new data, i.e. it should generalize. The fundamental works of Vapnik and further developments [16], [8], [5], show
that the key to obtain a meaningful solution to the above problem is to control the complexity of the solution space. Interestingly, as noted by [12], [8], [2], this is the idea underlying
regularization techniques for ill-posed inverse problems [15], [7]. In such a context to
avoid undesired oscillating behavior of the solution we have to restrict the solution space.
Not surprisingly the form of the algorithms proposed in both theories is strikingly similar.
Anyway a careful analysis shows that a rigorous connection between learning and regularization for inverse problem is not straightforward. In this paper we consider the square loss
and show that the problem of learning can be translated into a convenient inverse problem
and consistency results can be derived in a general setting. When a generic loss is considered the analysis becomes immediately more complicated.
Some previous works on this subject considered the special case in which the elements of
the input space are fixed and not probabilistically drawn [11], [9]. Some weaker results
in the same spirit of those presented in this paper can be found in [13] where anyway the
connections with inverse problems is not discussed. Finally, our analysis is close to the
idea of stochastic inverse problems discussed in [16]. It follows the plan of the paper. After recalling the main concepts and notation of learning and inverse problems, in section
4 we develop a formal connection between the two theories. In section 5 the main results
are stated and discussed. Finally in section 6 we conclude with some remarks and open
problems.
2
Learning from examples
We briefly recall some basic concepts of learning theory [16], [8]. In the framework of
learning, there are two sets of variables: the input space X, compact subset of R n , and
the output space Y , compact subset of R. The relation between the input x ? X and the
output y ? Y is described by a probability distribution ?(x, y) = ?(x)?(y|x) on X ? Y .
The distribution ? is known only through a sample z = (x, y) = ((x1 , y1 ), . . . , (x` , y` )),
called training set, drawn i.i.d. according to ?. The goal of learning is, given the sample
z, to find a function fz : X ? R such that fz (x) is an estimate of the output y when the
new input x is given. The function fz is called estimator and the rule that, given a sample
z, provides us with fz is called learning algorithm.
Given a measurable function f : X ? R, the ability of f to describe the distribution ? is
measured by its expected risk defined as
Z
I[f ] =
(f (x) ? y)2 d?(x, y).
X?Y
The regression function
g(x) =
Z
y d?(y|x),
Y
is the minimizer of the expected risk over the set of all measurable functions and always
exists since Y is compact. Usually, the regression function cannot be reconstructed exactly
since we are given only a finite, possibly small, set of examples z.
To overcome this problem, in the regularized least squares algorithm an hypothesis space
H is fixed, and, given ? > 0, an estimator fz ? is defined as the solution of the regularized
least squares problem,
`
min{
f ?H
1X
2
(f (xi ) ? yi )2 + ? kf kH }.
` i=1
(1)
The regularization parameter ? has to be chosen depending on the available data, ? =
?(`, z), in such a way that, for every > 0
?(`,z)
lim P I[fz
] ? inf I[f ] ? = 0.
(2)
`?+?
f ?H
We note that in general inf f ?H I[f ] is larger that I[g] and represents a sort of irreducible
error associated with the choice of the space H. The above convergence in probability is
usually called consistency of the algorithm [16] [14].
3
Ill-Posed Inverse Problems and Regularization
In this section we give a very brief account of linear inverse problems and regularization
theory [15], [7]. Let H and K be two Hilbert spaces and A : H ? K a linear bounded
operator. Consider the equation
Af = g?
(3)
where g? , g ? K and kg ? g? kK ? ?. Here g represents the exact, unknown data and g? the
available, noisy data. Finding the function f satisfying the above equation, given A and g ? ,
is the linear inverse problem associated to Eq. (3). The above problem is, in general, illposed, that is, the Uniqueness can be restored introducing the Moore-Penrose generalized
inverse f ? = A? g defined as the minimum norm solution of the problem
2
min kAf ? gkK .
f ?H
(4)
However the operator A? is usually not bounded so, in order to ensure a continuous dependence of the solution on the data, the following Tikhonov regularization scheme can be
considered1
2
2
min{kAf ? g? kK + ? kf kH },
(5)
f ?H
whose unique minimizer is given by
f?? = (A? A + ?I)?1 A? g? ,
(6)
where A? denotes the adjoint of A.
A crucial step in the above algorithm is the choice of the regularization parameter ? =
?(?, g? ), as a function of the noise level ? and the data g? , in such a way that
?(?,g? )
lim
f?
? f ?
= 0,
(7)
??0
H
?(?,g )
?
that is, the regularized solution f?
converges to the generalized solution f ? = A? g
?
(f exists if and only if P g ? Range(A), where P is the projection on the closure of the
range of A and, in that case, Af ? = P g) when the noise ? goes to zero.
The similarity between regularized least squares algorithm (1) and Tikhonov regularization (5) is apparent. However, several difficulties emerge. First, to treat the problem of
learning in the setting of ill-posed inverse problems we have to define a direct problem by
means of a suitable operator A. Second, in the context of learning, it is not clear the nature
of the noise ?. Finally we have to clarify the relation between consistency (2) and the kind
of convergence expressed by (7). In the following sections we will show a possible way to
tackle these problems.
4
Learning as an Inverse Problem
We can now show how the problem of learning can be rephrased in a framework close to
the one presented in the previous section.
We assume that hypothesis space H is a reproducing kernel Hilbert space [1] with a continuous kernel K : X ? X ? R. If x ? X, we let Kx (s) = K(s, x), and, if ? is the marginal
distribution of ? on X, we define the bounded linear operator A : H ? L2 (X, ?) as
(Af )(x) = hf, Kx iH = f (x),
1
In the framework of inverse problems, many other regularization procedures are introduced [7].
For simplicity we only treat the Tikhonov regularization.
that is, A is the canonical injection of H in L2 (X, ?). In particular, for all f ? H, the
expected risk becomes,
2
I[f ] = kAf ? gkL2 (X,?) + I[g],
where g is the regression function [2]. The above equation clarifies that if the expected
risk admits a minimizer fH on the hypothesis space H, then it is exactly the generalized
solution2 f ? = A? g of the problem
Af = g.
(8)
Moreover, given a training set z = (x, y), we get a discretized version Ax : H ? E` of A,
that is
(Ax f )i = hf, Kxi iH = f (xi ),
`
`
where E = R is the finite dimensional euclidean space endowed with the scalar product
`
1X
0
hy, y iE` =
yi yi0 .
` i=1
It is straightforward to check that
`
1X
2
(f (xi ) ? yi )2 = kAx f ? ykE` ,
` i=1
so that the estimator fz ? given by the regularized least squares algorithm is the regularized
solution of the discrete problem
Ax f = y.
(9)
At this point it is useful to remark the following two facts. First, in learning from examples
we are not interested into finding an approximation of the generalized solution of the discretized problem (9), but we want to find a stable approximation of the solution of the exact
problem (8) (compare with [9]). Second, we notice that in learning theory the consistency
property (2) involves the control of the quantity
2
2
I[fz ? ] ? inf I[f ] = kAf ? gkL2 (X,?) ? inf kAf ? gkL2 (X,?) .
f ?H
f ?H
If P is the projection on the closure of the range of A, the definition of P gives
2
I[fz ? ] ? inf I[f ] =
Afz ? ? P g
2
f ?H
L (X,?)
(10)
(11)
(the above equality stronlgy depends on the fact that the loss function is the square loss). In
the inverse problem setting, the square root of the above quantity is called the residue of the
solution fz ? . Hence, consistency
is
controlled by the residue of the estimator, instead of
?
?
the reconstruction error
fz ? f
(as in inverse problems). In particular, consistency
H
is a weaker condition than the one required by (7) and does not require the existence of the
generalized solution fH .
5
Regularization, Stochastic Noise and Consistency
To apply the framework of ill-posed inverse problems of Section 3 to the formulation of
learning proposed above, we note that the operator Ax in the discretized problem (9) differs
from the operator A in the exact problem (8) and a measure of the difference between A x
and A is required. Moreover, the noisy data y ? E` and the exact data g ? L2 (X, ?)
belong to different spaces, so that the notion of noise has to be modified. Given the above
premise our derivation of consistency results is developed in two steps: we first study the
residue of the solution by means of a measure of the noise due to discretization and then we
show a possible way to give a probabilistic evaluation of the noise previously introduced.
2
The fact that fH is the minimal norm solution of (4) is ensured by the assumption that the support
of the measure ? is X, since in this case the operator A is injective.
5.1
Bounding the Residue of the Regularized Solution
We recall that the regularized solutions of problems (9) and (8) are given by
fz?
=
(A?x Ax + ?I)?1 A?x y,
f?
=
(A? A + ?I)?1 A? g.
The above equations show that fz? and f ? depend only on A?x Ax and A? A which are
operators from H into H and on A?x y and A? g which are elements of H, so that the space
E` disappears. This observation suggests that noise levels could be kA?x Ax ? A? AkL(H)
and kA?x y ? A? gkH , where k?kL(H) is the uniform operator norm. To this purpose, for
every ? = (?1 , ?2 ) ? R2+ we define the collection of training sets.
U? = {z ? (X ? Y )` | kA?x y ? A? gkH ? ?1 , kA?x Ax ? A? AkL(H) ? ?2 , ` ? N}
and we let M = sup{|y| | y ? Y }. The next theorem is the central result of the paper.
Theorem 1 If ? > 0, the following inequalities hold
1. for any training set z ? U?
M?
?1
?
2
+ ?
Afz ? P g
L2 (X,?) ?
Af ? ? P g
L2 (X,?) ?
4?
2 ?
2. if P g ? Range(A), for any training set z ? U? ,
?
fz ? f ?
?
f ? ? f ?
? M ?32 + ?1
H
H
?
2? 2
Moreover if we choose ? = ?(?, z) in such a way that
?
?
? lim??0 supz?U? ?(?,2z)
?1
lim??0 supz?U? ?(?,z)
?
?2
? lim??0 sup
z?U?
?(?,z)
then
lim sup
Afz?(?,z) ? P g
??0 z?U?
= 0
= 0
= 0
L2 (X,?)
(12)
= 0.
(13)
We omit the complete proof and refer to [3]. Briefly, the idea is to note that
?
Afz ? P g
L2 (X,?) ?
Af ? ? P g
L2 (X,?)
1
?
Afz? ? Af ?
L2 (X,?) =
(A? A) 2 (fz? ? f ? )
H
where the last equation follows by polar decomposition of the operator A. Moreover a
simple algebraic computation gives
fz? ?f ? = (A? A+?I)?1 (A? A?A?x Ax )(A?x Ax +?I)?1 A?x y+(A? A+?I)?1 (A?x y?A? g)
where the relevant quantities for definition of the noise appear.
The first item in the above proposition quantifies the difference between the residues of
the regularized solutions of the exact and discretized problems in terms of the noise level
? = (?1 , ?2 ). As mentioned before this is exactly the kind of result needed to derive
consistency. On the other hand the last part of the proposition gives sufficient conditions
on the parameter ? to ensure convergence of the residue to zero as the level noise decreases.
The above results were obtained introducing the collection U? of training sets compatible
with a certain noise level ?. It is left to quantify the noise level corresponding to a training
set of cardinality `. This will be achieved in a probabilistic setting in the next section.
5.2
Stochastic Evaluation of the Noise
In this section we estimate the discretization noise ? = (?1 , ?2 ).
p
Theorem 2 Let 1 , 2 > 0 and ? = supx?X K(x, x), then
M?
?2
?
?
?
?
?
?
P kA g ? Ax ykH ?
+ 1 , kA A ? Ax Ax kL(H) ?
+ 2
`
`
2 `
1
2 `
2
? 1 ? e? 2?2 M 2 ? e? 2?4
(14)
The proof is given in [3] and it is based on McDiarmid inequality [10] applied to the random
variables
F (z) = kAx ? y ? A? gkH G(z) = kAx ? Ax ? A? AkL(H) .
Other estimates of the noise ? can be given using, for example, union bounds and Hoeffding?s inequality. Anyway rather then providing a tight analysis our concern was to find an
natural, explicit and easy to prove estimate of ?.
5.3
Consistency and Regularization Parameter Choice
Combining Theorems 1 and 2, we easily derive the following corollary.
Corollary 1 Given 0 < ? < 1, with probability greater that 1 ? ?,
?
Afz ? ? P g
? Af ? P g L2 (X,?)
2
L (X,?)
r
4
?M
1
?
? +
? ?
1 + log
?
2 `
? 2?
(15)
for all ? > 0.
Recalling (10) and (11) it is straightforward to check that the above inequality can be easily
restated in the usual learning notation, in fact we obtain
?2
?
r
?
? ?L
4
1
?
?
?
? +
I[fz ? ] ? ? ?
1 + log
+
Af ? ? P g
L2 (X,?) ? + inf I[f ] .
f ?H
?
?
?2 `
? 2?
|
{z
}
| {z }
|
{z
}
sample error
approximation error
irreducible error
In the above inequality the first term plays the role of sample error. If we choose the
regularization parameter so that ? = ?(`, z) = O( `1b ), with 0 < b < 12 the sample error
q
1
converges in probability to zero with order O
when ` ? ?. On the other
1?2b
`
hand the second term represents the approximation error and it is possible to show, using
standard results from spectral theory, that it vanishes as ? goes to zero [7]. Finally, the last
term represents the minimum attainable risk once the hypothesis space H has been chosen.
From the above observations it is clear that consistency is ensured once the parameter ? is
chosen according to the aforementioned conditions. Nonetheless to provide convergence
rates it is necessary to control the convergence rate of the approximation error. Unfortunately it is well known that this can be accomplished only making some assumptions on
the underlying probability distribution ? (see for example [2]). It can be shown that if the
explicit dependence of the approximation error on ? is not available we cannot determine
an optimal a priori (data independent) dependency ? = ?(`) for the regularization parameter. Nevertheless a posteriori (data dependent) choices ? = ?(`, z) can be considered to
automatically achieve optimal convergence rate [5], [6]. With respect to this last fact we
notice that the set of samples such that inequality (14) holds depends on ` and ?, but does
not depend ?, so that we can consider without any further effort a posteriori parameter
choices (compare with [4], [5]).
Finally, we notice that the estimate (15) is the result of two different procedures: Theorem 1, which is of functional type, gives the dependence of the bound by the regularization
parameter ? and by the noise levels kA?x Ax ? A? AkL(H) and kA?x y ? A? gkH , whereas
Theorem 2, which is of probabilistic nature, relates the noise levels to the number of data `
and the confidence level ?.
6
Conclusions
In this paper we defined a direct and inverse problem suitable for the learning problem and
derived consistency results for the regularized least squares algorithm. Though our analysis
formally explains the connections between learning theory and linear inverse problems, its
main limit is that we considered only the square loss. We briefly sketch how the arguments
presented in the paper extend to general loss functions. For sake of simplicity we consider
a differentiable loss function V . It is easy to see that the minimizer fH of the expected risk
satisfies the following equation
SfH = 0
(16)
where S = LK ? O and LK is the integral operator with kernel K, that is
Z
(LK f )(x) =
K(x, s)f (s)d?(s)
X
and O is the operator defined by
(Of )(x) =
Z
V 0 (y, f (x))d?(y|x).
Y
If we consider a generic differentiable loss the operator O and hence S is non linear, and
estimating fH is an ill-posed non linear inverse problem. It is well known that the theory
for this kind of problems is much less developed than the corresponding theory for linear
problems. Moreover, since, in general, I[f ] does not define a metric, it is not so clear the
relation between the expected risk and the residue. It appears evident that the attempt to
extend our results to a wider class of loss function is not straightforward. A possible way to
tackle the problem, further developing our analysis, might pass through the exploitation of
a natural convexity assumption on the loss function. Future work also aims to derive tighter
probabilistic bounds on the noise using recently proposed data dependent techniques.
Acknowledgments
We would like to thank M.Bertero, C. De Mol, M. Piana, T. Poggio, G. Talenti, A. Verri
for useful discussions and suggestions. This research has been partially funded by the
INFM Project MAIA, the FIRB Project ASTA2 and the IST Programme of the European
Community, under the PASCAL Network of Excellence, IST-2002-506778.
References
[1] N. Aronszajn. Theory of reproducing kernels. Trans. Amer. Math. Soc., 68:337?404,
1950.
[2] Felipe Cucker and Steve Smale. On the mathematical foundations of learning. Bull.
Amer. Math. Soc. (N.S.), 39(1):1?49 (electronic), 2002.
[3] E. De Vito, A. Caponnetto, and L. Rosasco.
Discretization error analysis
for Tikhonov regularization. submitted to Inverse Problem, 2004. available
http://www.disi.unige.it/person/RosascoL/publications/discre iop.pdf.
[4] E. De Vito, A. Caponnetto, and L. Rosasco. Model selection for regularized leastsquares algorithm in learning theory. to appear on Journal Machine Learning Research, 2004.
[5] L. Devroye, L. Gy?orfi, and G. Lugosi. A Probabilistic Theory of Pattern Recognition.
Number 31 in Applications of mathematics. Springer, New York, 1996.
[6] Schock E. and Sergei V. Pereverzev. On the adaptive selection of the parameter in
regularization of ill-posed problems. Technical report, University of Kaiserslautern,
august 200r.
[7] Heinz W. Engl, Martin Hanke, and Andreas Neubauer. Regularization of inverse problems, volume 375 of Mathematics and its Applications. Kluwer Academic Publishers
Group, Dordrecht, 1996.
[8] Theodoros Evgeniou, Massimiliano Pontil, and Tomaso Poggio. Regularization networks and support vector machines. Adv. Comput. Math., 13(1):1?50, 2000.
[9] Vera Kurkova. Supervised learning as an inverse problem. Technical Report 960,
Institute of Computer Science, Academy of Sciences of the Czech Republic, April
2004.
[10] Colin McDiarmid. On the method of bounded differences. In Surveys in combinatorics, 1989 (Norwich, 1989), volume 141 of London Math. Soc. Lecture Note Ser.,
pages 148?188. Cambridge Univ. Press, Cambridge, 1989.
[11] S. Mukherjee, T. Niyogi, P.and Poggio, and R. Rifkin. Statistical learning: Stability is
sufficient for generalization and necessary and sufficient for consistency of empirical
risk minimization. Technical Report CBCL Paper 223, Massachusetts Institute of
Technology, january revision 2004.
[12] T. Poggio and Girosi F. Networks for approximation and learning. Proc. IEEE,
78:1481?1497, 1990.
[13] Cynthia Rudin. A different type of convergence for statistical learning algorithms.
Technical report, Program in Applied and Computational Mathematics Princeton University, 2004.
[14] I. Steinwart. Consistency of support vector machines and other regularized kernel
machines. IEEE Transaction on Information Theory, 2004. (accepted).
[15] Andrey N. Tikhonov and Vasiliy Y. Arsenin. Solutions of ill-posed problems. V. H.
Winston & Sons, Washington, D.C.: John Wiley & Sons, New York, 1977. Translated
from the Russian, Preface by translation editor Fritz John, Scripta Series in Mathematics.
[16] Vladimir N. Vapnik. Statistical learning theory. Adaptive and Learning Systems
for Signal Processing, Communications, and Control. John Wiley & Sons Inc., New
York, 1998. A Wiley-Interscience Publication.
| 2722 |@word exploitation:1 version:1 briefly:3 norm:3 yi0:1 open:1 closure:2 decomposition:1 attainable:1 series:1 interestingly:1 ka:8 discretization:4 sergei:1 john:3 girosi:1 selected:1 rudin:1 item:1 provides:1 math:4 theodoros:1 mcdiarmid:2 mathematical:1 direct:2 prove:1 interscience:1 firb:1 excellence:1 expected:6 indeed:1 behavior:1 tomaso:1 nor:1 andrea:1 discretized:4 heinz:1 yke:1 automatically:1 considering:1 cardinality:1 becomes:2 project:2 estimating:1 underlying:2 notation:2 bounded:4 moreover:5 revision:1 vera:1 kg:1 kind:3 akl:4 developed:2 finding:2 disi:8 every:2 tackle:2 exactly:3 universit:5 ensured:2 ser:1 control:4 omit:1 appear:2 positive:1 before:1 treat:2 limit:1 lugosi:1 might:1 suggests:1 range:4 unique:1 acknowledgment:1 union:1 differs:1 illposed:1 procedure:2 pontil:1 empirical:1 orfi:1 convenient:1 projection:2 confidence:1 get:1 cannot:2 close:2 selection:2 operator:13 context:2 risk:8 www:1 measurable:2 pereverzev:1 straightforward:4 go:2 independently:1 survey:1 restated:1 simplicity:2 immediately:1 estimator:6 rule:1 supz:2 gkh:4 stability:1 anyway:3 notion:1 play:1 exact:5 hypothesis:4 element:2 satisfying:1 recognition:1 mukherjee:1 role:1 adv:1 decrease:1 mentioned:1 vanishes:1 convexity:1 complexity:1 vito:3 depend:2 tight:1 strikingly:1 translated:2 easily:2 derivation:1 univ:1 massimiliano:1 describe:1 london:1 odone:2 dordrecht:1 whose:1 unige:4 posed:9 larger:1 apparent:1 ability:1 niyogi:1 noisy:2 differentiable:2 reconstruction:1 product:1 relevant:1 combining:1 maia:1 rifkin:1 translate:1 achieve:1 academy:1 adjoint:1 kh:2 convergence:7 felipe:1 oscillating:1 converges:2 wider:1 depending:1 develop:1 derive:3 measured:1 eq:1 strong:1 soc:3 involves:1 quantify:1 stochastic:3 explains:1 require:1 premise:1 generalization:1 proposition:2 tighter:1 leastsquares:1 dipartimento:1 clarify:1 hold:2 considered:4 cbcl:1 fh:5 purpose:1 uniqueness:1 polar:1 proc:1 tool:1 minimization:1 always:1 aim:1 modified:1 rather:1 avoid:1 probabilistically:1 publication:2 corollary:2 derived:4 ax:15 check:2 rigorous:1 posteriori:2 dependent:2 relation:4 interested:1 aforementioned:1 ill:9 pascal:1 priori:1 development:1 plan:1 special:1 marginal:1 once:2 evgeniou:1 ernesto:1 washington:1 represents:4 future:1 report:4 irreducible:2 recalling:2 attempt:1 gkk:1 evaluation:2 integral:1 injective:1 necessary:2 poggio:4 euclidean:1 desired:1 theoretical:1 minimal:1 engl:1 bull:1 introducing:2 republic:1 subset:2 uniform:1 dependency:1 answer:1 supx:1 kxi:1 andrey:1 person:1 fritz:1 fundamental:1 ie:1 probabilistic:6 cucker:1 infn:1 central:1 rosasco:4 possibly:1 choose:2 hoeffding:1 account:1 de:5 gy:1 inc:1 combinatorics:1 umberto:2 depends:2 root:1 sup:3 sort:1 hf:2 complicated:1 hanke:1 square:10 clarifies:1 generalize:1 submitted:1 definition:2 nonetheless:1 associated:2 di:7 proof:2 massachusetts:1 recall:2 lim:6 hilbert:2 appears:1 steve:1 supervised:1 april:1 verri:1 formulation:1 amer:2 though:1 hand:2 sketch:1 steinwart:1 aronszajn:1 russian:1 concept:2 iop:1 regularization:24 equality:1 hence:2 moore:1 francesca:1 undesired:1 noted:1 generalized:5 pdf:1 evident:1 complete:1 recently:1 functional:1 volume:2 discussed:3 belong:1 extend:2 kluwer:1 refer:1 cambridge:2 consistency:15 mathematics:4 language:1 funded:1 stable:1 similarity:1 inf:6 tikhonov:5 certain:1 inequality:6 yi:3 accomplished:1 seen:1 minimum:2 greater:1 determine:1 colin:1 signal:1 relates:1 infer:1 caponnetto:4 technical:4 academic:1 af:9 controlled:1 kax:3 basic:1 regression:3 metric:1 kernel:5 achieved:1 whereas:1 want:1 residue:7 crucial:1 publisher:1 subject:1 spirit:1 easy:2 restrict:1 andreas:1 idea:3 effort:1 algebraic:1 york:3 remark:2 useful:2 clear:3 http:1 fz:17 canonical:1 notice:3 discrete:1 rephrased:1 ist:2 key:1 group:1 nevertheless:2 drawn:3 neither:1 inverse:26 kaiserslautern:1 electronic:1 genova:10 bound:3 winston:1 hy:1 sake:1 argument:1 min:3 injection:1 martin:1 developing:1 according:3 son:3 making:1 neubauer:1 equation:6 previously:1 needed:1 available:4 endowed:1 apply:1 generic:2 spectral:1 infm:1 sezione:1 existence:1 denotes:1 ensure:2 question:1 quantity:3 restored:1 dependence:3 usual:1 thank:1 devroye:1 kk:2 providing:1 vladimir:1 unfortunately:1 relate:1 smale:1 stated:1 unknown:2 perform:1 observation:2 finite:3 january:1 communication:1 y1:1 reproducing:2 august:1 community:1 introduced:2 required:2 kl:2 connection:5 czech:1 trans:1 usually:3 pattern:1 giovannini:1 program:1 suitable:2 difficulty:1 natural:2 regularized:12 scheme:1 technology:1 lorenzo:1 brief:1 disappears:1 lk:3 l2:11 kf:2 loss:11 lecture:1 suggestion:1 foundation:1 sufficient:3 scripta:1 editor:1 translation:1 compatible:1 arsenin:1 surprisingly:1 last:4 formal:2 weaker:2 institute:2 emerge:1 overcome:1 collection:2 adaptive:2 programme:1 matematica:1 transaction:1 reconstructed:1 compact:3 conclude:1 xi:3 piana:1 continuous:2 quantifies:1 nature:2 mol:1 european:1 main:4 bounding:1 noise:19 x1:1 wiley:3 explicit:2 comput:1 theorem:6 cynthia:1 r2:1 admits:1 evidence:1 concern:1 exists:2 ih:2 vapnik:2 kx:2 penrose:1 expressed:1 partially:1 scalar:1 springer:1 minimizer:4 satisfies:1 goal:2 careful:1 norwich:1 called:5 pas:1 accepted:1 meaningful:1 formally:1 support:3 princeton:1 |
1,897 | 2,723 | Resolving Perceptual Aliasing In The Presence
Of Noisy Sensors?
Ronen I. Brafman & Guy Shani
Department of Computer Science
Ben-Gurion University
Beer-Sheva 84105, Israel
{brafman, shanigu}@cs.bgu.ac.il
Abstract
Agents learning to act in a partially observable domain may need to
overcome the problem of perceptual aliasing ? i.e., different states that
appear similar but require different responses. This problem is exacerbated when the agent?s sensors are noisy, i.e., sensors may produce different observations in the same state. We show that many well-known
reinforcement learning methods designed to deal with perceptual aliasing, such as Utile Suffix Memory, finite size history windows, eligibility
traces, and memory bits, do not handle noisy sensors well. We suggest
a new algorithm, Noisy Utile Suffix Memory (NUSM), based on USM,
that uses a weighted classification of observed trajectories. We compare
NUSM to the above methods and show it to be more robust to noise.
1 Introduction
Consider an agent situated in a partially observable domain: It executes an action; the action
may change the state of the world; this change is reflected, in turn, by the agent?s sensors;
the action may have some associated cost, and the new state may have some associated
reward or penalty. Thus, the agent?s interaction with this environment is characterized by a
sequence of action-observation-reward steps, known as instances [7]. In this paper we are
interested in agents with imperfect and noisy sensors that learn to act in such environments
without any prior information about the underlying set of world-states and the world?s
dynamics, only information about their sensor?s capabilities. This is a known variant of
reinforcement learning (RL) in partially observable domains [1].
As the agent works with observations, rather than states, two possible problems arise: The
agent may observe too much data which requires computationally intensive filtering ? a
problem we do not discuss. Or, sensors may supply insufficient data to identify the current
state of the world based on the current observation. This leads to a phenomena known as
perceptual aliasing [2], where the same observation is obtained in distinct states requiring
different actions. For example, in Figure 1(a) the states marked with X are perceptually
aliased. Various RL techniques were developed to handle this problem.
?
Partially supported by the Lynn and William Frankel Center for Computer Sciences.
(a)
(b)
Figure 1: Two maze domains. If only wall configuration is sensed, states marked with the
same letter (X or Y) are perceptually aliased.
The problem of resolving the perceptual aliasing is exacerbated when the agent?s sensors
are not deterministic. For example, if walls can sometimes be reported where none exist, or
if a wall sometimes goes undetected. The performance of existing techniques for handling
RL domains with perceptually aliased states, such as finite size history windows, eligibility
traces, and internal memory quickly degrades as the level of noise increases.
In this paper, we introduce the Noisy Utile Suffix Memory (MUSM) algorithm, which
builds on McCallum?s Utile Suffix Memory (USM) algorithm [7]. We compare the performance of MUSM to USM and other existing methods on the above two standard maze
domains, and show that it is more robust to noise.
2 Background
We briefly review a number of algorithms for resolving the problem of perceptual aliasing.
We assume familiarity with basic RL techniques, and in particular, Q-learning, SARSA,
and eligibility traces (see Sutton and Barto [12] for more details).
The simplest way to handle perceptual aliasing is to ignore it by using the observation
space as the state space, defining a memory-less policy. This approach works sometimes,
although it generally results in poor performance. Jaakkola et al. [4] suggest stochastic
memory-less policies, and Williams and Singh [13] implement an online version of their
algorithm and showed that it approximates an optimal solution with 50% accuracy.
Eligibility traces can be viewed as a type of short-term memory, as they update the recently
visited state-action couplings. Therefore, they can be used to augment the ability to handle
partial observability. Loch and Singh [6] explore problems where a memory-less optimal
policy exists and demonstrated that Sarsa(?) can learn an optimal policy for such domains.
Finite-size history methods are a natural extension to memory-less policies. Instead of
identifying a state with the last observation, we can use the last k observations. The size of
the history window (k) is a fixed parameter, identical for all observation sequences. Lock
and Singh [6] show Sarsa(?) using fixed history window to learn a good policy in domains
where short-term memory optimal policies exist.
An arbitrarily predefined history window cannot generally solve perceptual aliasing: an
agent cannot be expected to know in advance how long should it remember the actionobservation-reward trajectories. Usually, some areas of the state space require the agent to
remember more, while in other locations a reactive policy is sufficient. A better solution is
to learn online the history length needed to decide on the best action in the current location.
McCallum [7] extensively handles those issues in his dissertation. We review McCallums?
Utile Suffix Memory (USM) algorithm in Section 3.
Another possible approach to handle perceptual aliasing is to augment the observations
with some internal memory [10]. The agents? state s is composed of both the current
observation o and the internal memory state m. The agents? actions are enhanced with
actions that modify the memory state by flipping one of the bits. The agent uses some
standard learning mechanism to learn the proper action, including actions that change the
memory state. The algorithm is modified so that no cost or decay is applied to the actions
that modify the memory. This approach is better than using finite or variable history length
because meaningful events can occur arbitrarily far in the past. Keeping all the possible
trajectories from the event onwards until the outcome is observed might cost too much, and
McCallums? techniques are unable to group those trajectories together to deduce the proper
action. Peshkin et al. [10] demonstrated the memory-bits approach to converge, but did not
show it to be superior to any other algorithm.
Other approaches include the use of finite-state automata (FSA) [8] (which can be viewed
as a special case of the memory-bits approach); the use of neural networks for internal
memory [5, 3]; and constructing and solving a POMDP model [7, 2]. The new emerging
technique of Predictive State Representations (PSR) [11] may also provide a good way
to online learn a model of the environment. We note that most researchers tested their
algorithms on environments with very little noise, and do not examine the effect of noise
on their performance.
3 Utile Suffix Memory
Instance-based state identification [7] resolves perceptual aliasing with variable length short
term memory. An Instance is a tuple Tt = hTt?1 , at?1 , ot , rt i ? the individual observed
raw experience. Algorithms of this family keep all the observed raw data (sequences of
instances), and use it to identify matching subsequences. The algorithm assumes that if the
suffix of two sequences is similar both were likely generated in the same world state.
Utile Suffix Memory creates a tree structure, based on the well known suffix trees for string
operations. This tree maintains the raw experiences and identifies matching suffixes. The
root of the tree is an unlabelled node, holding all available instances. Each immediate child
of the root is labelled with one of the observations encountered during the test. A node
holds all the instances Tt = hTt?1 , at?1 , ot , rt i whose final observation ot matches the
observation in the node?s label. At the next level, instances are split based on the last action
of the instance at . Then, we split again based on (the next to last) observation ot?1 , etc.
All nodes act as buckets, grouping together instances that have matching history suffixes of
a certain length. Leaves take the role of states, holding Q-values and updating them. The
deeper a leaf is in the tree, the more history the instances in this leaf share.
The tree is built on-line during the test run. To add a new instance to the tree, we examine
its precept, and follow the path to the child node labeled by that precept. We then look at
the action before this precept and move to the node labeled by that action, then branch on
the precept prior to that action and so forth, until a leaf is reached.
Identifying the proper depth for a certain leaf is a major issue, and we shall present a number of improvements to McCallum?s methods. Leaves should be split if their descendants
show a statistical difference in expected future discounted reward associated with the same
action. We split instances in a node if knowing where the agent came from helps predict
future discounted rewards. Thus, the tree must keep what McCallum calls fringes, i.e.,
subtrees below the ?official? leaves.
For better performance, McCallum did not compare the nodes in the fringes to their siblings, only to their ancestor ?official? leaf. He also did not compare values from all actions
executed from the fringe, only the action that has the highest Q-value in the leaf (the policy
action of that leaf). To compare the populations of expected discounted future rewards from
the two nodes (the fringe and the ?official? leaf), he used the Kolmogorov-Smirnov (KS)
test ? a non-parametric statistical test used to find whether two populations were generated by the same distribution. If the test reported that a difference was found between the
expected discounted future rewards after executing the policy action, the leaf was split, the
fringe node would become the new leaf, and the tree would be expanded to create deeper
fringes. Figure 3 presents an example of a possible USM tree, without fringe nodes.
Figure 2: A possible USM suffix tree generated by the maze in Figure 1(a). Below is a
sequence of instances demonstrating how some instances are clustered into the tree leaves.
Instead of comparing the fringe node to its ancestor ? official? leaf, we found it computationally possible to compare the siblings of the fringe, avoiding the problem that the same
instance appears in both distributions. McCallum compared only the expected discounted
future rewards from executing the policy action, where we compare all the values following
all actions executed after any of the instances in the fringe. McCallum used the KS test,
where we choose to use the more robust randomization test [14] that works well with small
sets of instances. McCallum also considered only fringe nodes of certain depth, given as
a parameter to the algorithm, where we choose to create fringe nodes as deep as possible,
until the number of instances in the node diminish below some threshold (we use a value
of 10 in our experiments).
The expected future discounted reward of instance Ti is defined by:
Q(Ti ) = ri + ?U (L(Ti+1 ))
(1)
where L(Ti ) is the leaf associated with instance Ti and U (s) = maxa (Q(s, a)).
After inserting new instances to the tree, we update Q-values in the leaves using:
P
Ti ?T (s,a) ri
R(s, a) =
|T (s, a)|
|?Ti ? T (s, a), L(Ti+1 ) = s0 |
P r(s0 |s, a) =
|T (s, a)|
X
P r(s0 |s, a)U (s0 )
Q(s, a) = R(s, a) + ?
(2)
(3)
(4)
s0
We use s and s0 to denote the leaves of the tree, as in an optimal tree configuration for a
problem the leaves of the tree define the sates of the underlying MDP. The above equations
therefore correspond to a single step of the value iteration algorithm used in MDPs.
Now that the Q-values have been updated, the agent chooses the next action to perform
based on the Q-values in the leaf corresponding to the current instance Tt :
at+1 = argmaxa Q(L(Tt ), a)
(5)
McCallum uses the fringes of the tree for a smart exploration strategy. In our implementation we use a simple ?-greedy technique for exploration.
4 Adding Noise to Utile Suffix Memory
There are two types of noise in perception. The system makes different observations at the
same location (false negative), or it makes identical observations at different locations (false
positive). USM handles false positives by differentiating identical current observations
using the agents? history. Knowing that the agent came from different locations, helps it
realize that it is in two different locations, though the observations look the same.
USM does not handle false negative perceptions well. When the agent is at the same state,
but receives different observations, it is unable to learn from the noisy observation and thus
wastes much of the available information. Our Noisy Utile Suffix Memory (NUSM) is
designed to overcome this weakness.
It is reasonable to assume that the agent has some sensor model defining pr(o|s) ? the
probability that the agent will observe o in world state s. We can use the sensor model to
augment USM with the ability to learn from noisy instances. In our experiments we assume
the agent has n boolean sensors with an accuracy probability pi . A single observation is
composed of n output values o = h?1 , ..., ?n i when ?i ? 0, 1. The probability that an
observation o = h?1o , ..., ?1o i came from an actual world state s = h?1s , ..., ?ns i is therefore:
pr(o|s) =
n
Y
(6)
?i
i=0
?i =
?
pi
(1 ? pi )
?i0 = ?is
?i0 6= ?is
(7)
USM inserts each instance into a single path, ending at one leaf. Using any sensor model
we can insert the new instance Tt = hTt?1 , at?1 , ot , rt i into several paths with different
weights. When inserting Tt with weight w into an action node at depth k (with its children
labeled by observations) we will insert the instance into every child node c, with weight
w ? pr(ot?k?1 |label(c)). When inserting Tt with weight w into an observation node at
depth k (with its children labeled by actions) we will insert the instance only into the child
c labeled by at?k?1 with the same weight w. Weights of instances are stored in each node
with the instance as ws (Tt ) ? the weight of instance Tt in node s.
We can now rewrite Equation 2 and Equation 3:
P
R(s, a) =
0
T ?T (s,a) ri
Pi
P r(s |s, a) =
Ti ?T (s,a)
P
? ws (Ti )
ws (Ti )
T ,L(Ti+1 )=s0
Pi
Ti ?T (s,a)
ws (Ti )
ws (Ti )
(8)
(9)
The noisy instances are used only for updating the Q-values in the leaves. The test for
splitting is still calculated using identical sequences only. The tree structure of USM and
NUSM is hence identical, but NUSM learns a better policy. Conducted experiments indicate that using the noisy sequences for deciding when to split leaves provides a slight gain
in collected rewards, but the constructed tree is much larger, resulting in a considerable hit
to performance.
NUSM learns noisy sequences better. When a state corresponding to a noisy sequence is
observed, even though the noise in it might make it rare, NUSM can still use data from real
sequences to decide which action is useful for this state.
5 Experimental Results
To test our algorithms we used two maze worlds, seen in Figure 1(a) and Figure 1(b),
identical to the worlds McCallum used to show the performance of the USM algorithm. In
both cases some of the world states are perceptually aliased and the algorithm should learn
to identify the real world states. The agent in our experiments has four sensors allowing it
to sense an immediate wall above, below, to the left, and to the right of its current location.
Sensors have a boolean output that has a probability ? of being accurate. The probability
of all sensors providing the correct output is ?4 . In both mazes there is a single location
that grants the agent a reward of 10. Upon receiving that reward the agent is transformed to
any of the perceptually aliased states of the maze randomly. If the agent bumps into a wall
it pays a cost (a negative reward) of 1. For every move the agent pays a cost of 0.1.
We compared the performance of applying finite size history windows to Q-learning and
Sarsa, eligibility traces, memory bits, USM and NUSM on the two worlds. In the tables
below, QL2 denotes using Q-learning with a history window of size 2, and S2 denotes using
Sarsa with a window size of 2. S(?) denotes the Sarsa(?) algorithm. Adding the superfix 1
denotes adding 1 bit of internal memory. For example, S(?)12 denotes using Sarsa(?) with a
history window of 2 and 1 internal memory bit. The columns and rows marked ? 2 present
the average variance over methods (bottom row) and ? values (rightmost column). In the
NUSM column, in brackets, is the improvement NUSM gains over USM.
As we are only interested in the effect of noisy sensors, the maze examples we use do
not demonstrate the advantages of the various algorithms; USM?s ability to automatically
compute the needed trajectory length in different locations and the internal memory ability
to remember events that occurred arbitrary far in the past are unneeded since our examples
require the agent to look only at the last 2 instances in every perceptually aliased location.
We ran each algorithm for 50000 steps learning a policy as explained above, and calculated
the average reward over the last 5000 iterations only to avoid the difference in convergence
time. We ran experiments with varying values of ? (accuracy of the sensors) ranging from
1.00 (sensor output is without noise) to 0.85 (overall output accuracy of 0.522). Reported
results are averaged over 10 different executions of each algorithm. We also ran experiments for Sarsa and Q-learning with only the immediate observation, which yielded poor
results as expected, and for history window of size 3 and 4 which resulted in lower performance than history window of size 2 for all algorithms (and were therefore removed from
the tables). Additional memory bits did not improve performance either. In our Sarsa(?)
implementation we used ? = 0.9.
?
QL2
S2
S(?)1
S(?)2
S(?)3
S(?)11
S(?)12
USM
NUSM
?2
1.00
0.99
0.98
0.97
0.96
0.95
0.94
0.93
0.92
0.91
0.90
0.89
0.88
0.87
0.86
0.85
?2
1.51
1.46
1.42
1.36
1.28
1.24
1.11
1.05
0.96
0.88
0.83
0.74
0.61
0.58
0.40
0.45
0.01
1.54
1.47
1.42
1.38
1.26
1.23
1.10
1.03
0.88
0.82
0.73
0.60
0.59
0.50
0.37
0.35
0.015
0.32
0.33
0.32
0.41
0.38
0.43
0.47
0.42
0.47
0.47
0.46
0.48
0.42
0.48
0.46
0.45
0.004
1.53
1.47
1.36
1.31
1.29
1.21
1.16
1.13
1.10
1.06
1.02
0.95
0.90
0.85
0.79
0.75
0.004
0.98
0.98
0.83
0.64
0.58
0.55
0.45
0.43
0.33
0.29
0.22
0.23
0.14
0.12
0.07
0.05
0.022
1.27
1.34
1.41
1.35
1.24
1.26
0.89
0.94
0.92
0.74
0.77
0.80
0.66
0.63
0.55
0.47
0.061
1.53
1.45
1.40
1.35
1.27
1.22
1.12
1.07
1.04
0.94
0.92
0.87
0.83
0.78
0.76
0.68
0.005
1.56
1.49
1.42
1.38
1.35
1.30
1.18
1.16
1.12
0.96
0.99
0.93
0.84
0.71
0.57
0.47
0.02
1.57(+0%)
1.54(+3%)
1.44(+1%)
1.43(+4%)
1.40(+4%)
1.35(+4%)
1.29(+9%)
1.29(+11%)
1.20(+7%)
1.12(+17%)
1.07(+8%)
1.04(+12%)
1.01(+20%)
0.98(+38%)
0.92(+61%)
0.87(+85%)
0.003
0.033
0.016
0.008
0.007
0.014
0.009
0.025
0.023
0.014
0.022
0.013
0.015
0.013
0.011
0.021
0.018
Table 1: Average reward as function of sensor accuracy, for the maze in Figure 1(a).
?
QL2
S2
S(?)1
S(?)2
S(?)3
S(?)11
S(?)12
USM
NUSM
?2
1.00
0.99
0.98
0.97
0.96
0.95
0.94
0.93
0.92
0.91
0.90
0.89
0.87
0.86
0.85
?2
1.42
1.40
1.33
1.30
1.26
1.19
1.09
1.05
0.94
0.85
0.69
0.64
0.44
0.30
0.08
0.008
1.46
1.41
1.35
1.29
1.25
1.16
1.05
0.94
0.84
0.80
0.72
0.58
0.42
0.21
0.10
0.01
0.23
0.24
0.25
0.26
0.24
0.21
0.14
0.12
0.12
0.09
0.10
0.06
0.06
0.04
0.01
0.003
1.53
1.44
1.35
1.22
1.06
1.00
0.93
0.82
0.72
0.77
0.65
0.58
0.47
0.26
0.29
0.009
1.49
1.34
1.24
1.15
1.06
0.90
0.85
0.76
0.64
0.47
0.42
0.24
0.16
0.09
0.09
0.006
1.47
1.24
0.94
0.90
0.43
0.33
0.30
0.39
0.30
0.24
0.23
0.24
0.16
0.13
0.18
0.071
1.54
1.43
1.34
1.21
1.12
1.03
0.93
0.77
0.66
0.48
0.48
0.40
0.31
0.20
0.16
0.016
1.75
1.57
1.43
1.40
1.28
1.23
1.09
1.09
1.02
0.93
0.87
0.81
0.72
0.68
0.61
0.011
1.72(-2%)
1.61(+3%)
1.46(+2%)
1.40(0%)
1.31(+2%)
1.26(+2%)
1.14(+5%)
1.09(+0%)
1.03(+1%)
0.96(+3%)
0.91(+5%)
0.92(+14%)
0.82(+14%)
0.81(+19%)
0.75(+23%)
0.006
0.004
0.027
0.034
0.032
0.015
0.015
0.011
0.018
0.011
0.013
0.013
0.010
0.012
0.015
0.008
Table 2: Average reward as function of sensor accuracy, for the maze in Figure 1(b).
4/
6/0%
$YHUDJHUHZDUG
6/
860
1860
,QVWDQFHV
Figure 3: Convergence rates for the maze in Figure 1(a) when sensor accuracy is 0.9.
As we can see, when sensor output is only slightly noisy, all algorithms perform reasonably;
NUSM performs the best, but differences are minor. This is because when sensors supply
perfect output, resolving the perceptual aliasing results in a fully observable environment
which Q-learning and Sarsa can solve optimally. When noise increases, the ability of
NUSM to use similar suffixes of trajectories, results in a noticeable performance gain over
other algorithms. The only algorithm that competes with NUSM is Sarsa(?) with a history
window of 2. The ability of Sarsa(?) to perform well in partially observable domains have
been noted by Lock and Singh [6]1 , but we note here that the performance of Sarsa(?) relies
heavily on the proper definition of the required history window size. When the history
window differs slightly from the required size, the performance hit is substantial, as we can
see in the two adjacent columns.
NUSM is more expensive computationally than USM and takes longer to converge. In
Figure 3, we can see that it still converges reasonably fast. Moreover, each NUSM iteration
takes about 5.6 milliseconds, when a USM iteration takes 3.1 milliseconds with the same
1
Lock and Singh also recommend the use of replacing traces but we found that using accumulating
traces produced better performance.
accuracy of ? = 0.85 and a similar number of nodes (10, 229 for NUSM and 10, 349 for
USM ? including fringe nodes), making NUSM reasonable for online learning.
Finally, Both USM and NUSM attempt to disambiguate the perceptual aliasing and create
a fully observable MDP. Yet, it is better to model the world directly as partially observable
using a Partially Observable Markov Decision Process (POMDP). POMDP policies explicitly address the problem of incomplete knowledge, taking into account the ability of certain
actions to reduce uncertainty without immediately generating useful rewards. Nikovski [9]
used McCallum?s Nearest Sequence Memory (NSM), a predecessor of USM to generate
and solve a POMDP from the observed instances. They, however, considered environments with little noise. We implemented their algorithms and obtained poor results in the
presence of noise in our domains, probably due to the use of NSM for state identification.
6 Conclusions
We show that some RL algorithms, including finite size history windows, memory bits
and USM, that resolve perceptual aliasing, provide poor results in the presence of noisy
sensors. We provided some insights as to why McCallums? USM algorithm does not handle
well noisy input from the agents? sensors and introduce NUSM ? an extension to USM
that learns from noisy sequences and handles environments where sensors provide noisy
output better. As noise arises, NUSM works better than other algorithms used for handling
domains with perceptual aliasing.
References
[1] A. R. Cassandra, L. P. Kaelbling, and M. L. Littman. Acting optimally in partially observable
stochastic domains. In AAAI?94, pages 1023?1028, 1994.
[2] L. Chrisman. Reinforcement learning with perceptual aliasing: The perceptual distinctions
approach. In AAAI?02, pages 183?188, 1992.
[3] S. Hochreiter and J. Schmidhuber. Long short-term memory. Neural Computation, 9(8):1735?
1780, 1997.
[4] T. Jaakkola, S. P. Singh, and M. I. Jordan. Reinforcement learning algorithm for partially
observable Markov decision problems. In NIPS?95, volume 7, pages 345?352, 1995.
[5] L.-J. Lin and T. M. Mitchell. Memory approaches to reinforcement learning in non-markovian
domains. Technical Report CMU-CS-92-138, 1992.
[6] J. Loch and S. Singh. Using eligibility traces to find the best memoryless policy in partially
observable Markov decision processes. In ICML?98, pages 323?331, 1998.
[7] A. K. McCallum. Reinforcement Learning with Selective Perception and Hidden State. PhD
thesis, University of Rochester, 1996.
[8] N. Meuleau, L. Peshkin, K. Kim, and L. P. Kaelbling. Learning finite-state controllers for
partially observable environments. In UAI?99, pages 427?436, 1999.
[9] D. Nikovski. State-Aggregation Algorithms for Learning Probabilistic Models for Robot Control. PhD thesis, Carnegie Mellon University, 2002.
[10] L. Peshkin, N. Meuleau, and L. P. Kaelbling. Learning policies with external memory. In
ICML?99, pages 307?314, 1999.
[11] S. Singh, M. L. Littman, and R. S. Sutton. Predictive representations of state. In NIPS 2001,
pages 1555?1561, December 2001.
[12] R. S. Sutton and A. G. Barto. Reinforcement Learning: An Introduction. MIT Press, 1998.
[13] J. K. Williams and S. Singh. Experimental results on learning stochastic memoryless policies
for partially observable markov decision processes. In NIPS, 1998.
[14] A. Yeh. More accurate tests for the statistical significance of result differences. In 18th Int.
Conf. on Computational Linguistics, pages 947?953, 2000.
| 2723 |@word version:1 briefly:1 smirnov:1 sensed:1 configuration:2 rightmost:1 past:2 existing:2 current:7 comparing:1 yet:1 must:1 realize:1 gurion:1 designed:2 update:2 greedy:1 leaf:23 mccallum:12 short:4 dissertation:1 meuleau:2 utile:9 provides:1 node:22 location:10 constructed:1 become:1 supply:2 predecessor:1 descendant:1 introduce:2 expected:7 examine:2 aliasing:15 discounted:6 automatically:1 resolve:2 little:2 actual:1 window:15 provided:1 underlying:2 competes:1 moreover:1 aliased:6 israel:1 what:1 string:1 emerging:1 maxa:1 developed:1 shani:1 remember:3 every:3 act:3 ti:15 hit:2 control:1 grant:1 appear:1 before:1 positive:2 modify:2 sutton:3 path:3 might:2 k:2 averaged:1 implement:1 differs:1 area:1 matching:3 argmaxa:1 suggest:2 cannot:2 applying:1 accumulating:1 deterministic:1 demonstrated:2 center:1 go:1 williams:2 automaton:1 pomdp:4 identifying:2 splitting:1 immediately:1 insight:1 his:1 population:2 handle:10 updated:1 enhanced:1 heavily:1 us:3 expensive:1 updating:2 labeled:5 observed:6 role:1 bottom:1 highest:1 removed:1 ran:3 substantial:1 environment:8 reward:17 littman:2 dynamic:1 singh:9 solving:1 rewrite:1 smart:1 predictive:2 creates:1 upon:1 various:2 kolmogorov:1 distinct:1 fast:1 outcome:1 whose:1 larger:1 solve:3 ability:7 noisy:19 final:1 online:4 fsa:1 sequence:12 precept:4 advantage:1 interaction:1 inserting:3 forth:1 nikovski:2 convergence:2 produce:1 generating:1 perfect:1 executing:2 ben:1 converges:1 help:2 coupling:1 ac:1 nearest:1 minor:1 noticeable:1 exacerbated:2 implemented:1 c:2 indicate:1 correct:1 stochastic:3 exploration:2 require:3 clustered:1 wall:5 randomization:1 sarsa:13 extension:2 insert:4 hold:1 considered:2 diminish:1 deciding:1 predict:1 bump:1 major:1 label:2 visited:1 create:3 weighted:1 mit:1 sensor:27 modified:1 rather:1 avoid:1 varying:1 barto:2 jaakkola:2 improvement:2 kim:1 sense:1 suffix:15 i0:2 unneeded:1 w:5 hidden:1 ancestor:2 transformed:1 selective:1 interested:2 issue:2 classification:1 overall:1 augment:3 special:1 identical:6 look:3 icml:2 future:6 report:1 recommend:1 randomly:1 nsm:2 composed:2 resulted:1 individual:1 william:1 attempt:1 onwards:1 weakness:1 bracket:1 predefined:1 subtrees:1 accurate:2 tuple:1 partial:1 experience:2 tree:19 incomplete:1 instance:33 column:4 boolean:2 markovian:1 cost:5 kaelbling:3 rare:1 conducted:1 too:2 optimally:2 reported:3 actionobservation:1 stored:1 chooses:1 probabilistic:1 receiving:1 together:2 quickly:1 again:1 aaai:2 thesis:2 choose:2 guy:1 bgu:1 external:1 conf:1 account:1 waste:1 int:1 explicitly:1 root:2 reached:1 aggregation:1 maintains:1 capability:1 rochester:1 il:1 accuracy:8 variance:1 correspond:1 identify:3 ronen:1 identification:2 raw:3 produced:1 none:1 trajectory:6 researcher:1 executes:1 history:20 definition:1 associated:4 gain:3 mitchell:1 knowledge:1 appears:1 follow:1 reflected:1 response:1 though:2 until:3 receives:1 replacing:1 mdp:2 effect:2 requiring:1 hence:1 memoryless:2 deal:1 adjacent:1 during:2 eligibility:6 noted:1 tt:9 demonstrate:1 performs:1 ranging:1 recently:1 superior:1 rl:5 volume:1 he:2 approximates:1 slight:1 occurred:1 mellon:1 robot:1 longer:1 deduce:1 etc:1 add:1 showed:1 schmidhuber:1 certain:4 arbitrarily:2 came:3 frankel:1 seen:1 additional:1 converge:2 resolving:4 branch:1 technical:1 match:1 unlabelled:1 characterized:1 long:2 lin:1 variant:1 basic:1 controller:1 cmu:1 iteration:4 sometimes:3 hochreiter:1 background:1 ot:6 probably:1 december:1 jordan:1 call:1 presence:3 htt:3 split:6 imperfect:1 observability:1 reduce:1 knowing:2 sibling:2 intensive:1 peshkin:3 whether:1 sheva:1 penalty:1 action:29 deep:1 generally:2 useful:2 extensively:1 situated:1 simplest:1 generate:1 exist:2 millisecond:2 carnegie:1 shall:1 group:1 four:1 demonstrating:1 threshold:1 run:1 letter:1 uncertainty:1 family:1 reasonable:2 decide:2 decision:4 bit:9 pay:2 encountered:1 yielded:1 occur:1 ri:3 expanded:1 department:1 poor:4 slightly:2 making:1 explained:1 pr:3 bucket:1 computationally:3 equation:3 turn:1 discus:1 mechanism:1 needed:2 know:1 available:2 operation:1 observe:2 assumes:1 denotes:5 include:1 linguistics:1 lock:3 build:1 move:2 flipping:1 degrades:1 parametric:1 rt:3 strategy:1 unable:2 collected:1 loch:2 length:5 insufficient:1 providing:1 executed:2 lynn:1 holding:2 trace:8 negative:3 implementation:2 proper:4 policy:17 perform:3 allowing:1 observation:26 markov:4 finite:8 immediate:3 defining:2 arbitrary:1 required:2 distinction:1 chrisman:1 nip:3 address:1 usually:1 below:5 perception:3 built:1 including:3 memory:36 event:3 natural:1 undetected:1 improve:1 mdps:1 identifies:1 prior:2 review:2 yeh:1 fully:2 filtering:1 agent:29 sufficient:1 beer:1 s0:7 share:1 pi:5 row:2 brafman:2 supported:1 keeping:1 last:6 deeper:2 taking:1 differentiating:1 overcome:2 depth:4 calculated:2 world:13 ending:1 maze:10 reinforcement:7 far:2 observable:13 ignore:1 usm:25 keep:2 uai:1 subsequence:1 why:1 table:4 disambiguate:1 learn:9 reasonably:2 robust:3 constructing:1 domain:13 official:4 did:4 significance:1 s2:3 noise:13 arise:1 child:6 n:1 perceptual:16 learns:3 familiarity:1 decay:1 grouping:1 exists:1 false:4 adding:3 phd:2 execution:1 perceptually:6 cassandra:1 explore:1 psr:1 likely:1 partially:12 relies:1 fringe:14 marked:3 viewed:2 labelled:1 considerable:1 change:3 acting:1 experimental:2 meaningful:1 internal:7 arises:1 reactive:1 phenomenon:1 tested:1 avoiding:1 handling:2 |
1,898 | 2,724 | Kernel Methods for Implicit Surface Modeling
?
Bernhard Sch?olkopf? , Joachim Giesen+? & Simon Spalinger+
Max Planck Institute for Biological Cybernetics, 72076 Tu? bingen, Germany
[email protected]
+
Department of Computer Science, ETH Zu? rich, Switzerland
[email protected],[email protected]
Abstract
We describe methods for computing an implicit model of a hypersurface
that is given only by a finite sampling. The methods work by mapping
the sample points into a reproducing kernel Hilbert space and then determining regions in terms of hyperplanes.
1
Introduction
Suppose we are given a finite sampling (in machine learning terms, training data)
x1 , . . . , xm ? X , where the domain X is some hypersurface in Euclidean space Rd . The
case d = 3 is especially interesting since these days there are many devices, e.g., laser
range scanners, that allow the acquisition of point data from the boundary surfaces of
solids. For further processing it is often necessary to transform this data into a continuous model. Today the most popular approach is to add connectivity information to the data
by transforming them into a triangle mesh (see [4] for an example of such a transformation
algorithm). But recently also implicit models, where the surface is modeled as the zero set
of some sufficiently smooth function, gained some popularity [1]. They bear resemblance
to level set methods used in computer vision [6]. One advantage of implicit models is that
they easily allow the derivation of higher order differential quantities such as curvatures.
Another advantage is that an inside-outside test, i.e., testing whether a query point lies on
the bounded or unbounded side of the surface, boils down to determining the sign of a
function-evaluation at the query point. Inside-outside tests are important when one wants
to intersect two solids.
The goal of this paper is, loosely speaking, to find a function which takes the value zero on
a surface which
(1) contains the training data and
(2) is a ?reasonable? implicit model of X .
To capture properties of its shape even in the above general case, we need to exploit some
structure on X . In line with a sizeable amount of recent work on kernel methods [11], we
assume that this structure is given by a (positive definite) kernel, i.e., a real valued function
?
Partially supported by the Swiss National Science Foundation under the project ?Non-linear
manifold learning?.
o
Figure 1: In the 2-D toy example depicted,
the hyperplane hw, ?(x)i = ? separates all
but one of the points from the origin. The outlier ?(x) is associated with a slack variable ?,
which is penalized in the objective function
(4). The distance from the outlier to the hyperplane is ?/kwk; the distance between hyperplane and origin is ?/kwk. The latter implies that a small kwk corresponds to a large
margin of separation from the origin.
o
o
o
.
o
o
o
?/ ||w||
w
? (x) o
o
?/ ||w||
o
k on X ? X which can be expressed as
k(x, x0 ) = h?(x), ?(x0 )i
(1)
for some map ? into a Hilbert space H. The space H is the reproducing kernel Hilbert
space (RKHS) associated with k, and ? is called its feature map. A popular example, in
the case where X is a normed space, is the Gaussian (where ? > 0)
kx ? x0 k2
0
.
(2)
k(x, x ) = exp ?
2 ?2
The advantage of using a positive definite kernel as a similarity measure is that it allows us
to construct geometric algorithms in Hilbert spaces.
2
Single-Class SVMs
Single-class SVMs were introduced [8, 10] to estimate quantiles C ? {x ? X |f (x) ?
[?, ?[} of an unknown distribution P on X using kernel expansions. Here,
X
f (x) =
?i k(xi , x) ? ?,
(3)
i
where x1 , . . . , xm ? X are unlabeled data generated i.i.d. according to P . The single-class
SVM approximately computes the smallest set C ? C containing a specified fraction of
all training examples, where smallness is measured in terms of the norm in the RKHS H
associated with k, and C is the family of sets corresponding to half-spaces in H. Depending
on the kernel, this notion of smallness will coincide with the intuitive idea that the quantile
estimate should not only contain a specified fraction of the training points, but it should
also be sufficiently smooth so that the same is approximately true for previously unseen
points sampled from P .
Let us briefly describe the main ideas of the approach. The training points are mapped into
H using the feature map ? associated with k, and then it is attempted to separate them from
the origin with a large margin by solving the following quadratic program: for ? ? (0, 1], 1
1
1 X
minimize
kwk2 +
?i ? ?
(4)
2
?m i
w?H,??Rm ,??R
subject to
hw, ?(xi )i ? ? ? ?i , ?i ? 0.
(5)
Since non-zero slack variables ?i are penalized in the objective function, we can expect that
if w and ? solve this problem, then the decision function, f (x) = sgn (hw, ?(x)i ? ?) will
1
Here and below, bold face greek character denote vectors, e.g., ? = (? 1 , . . . , ?m )> , and indices
i, j by default run over 1, . . . , m.
Figure 2: Models computed with a single class SVM using a Gaussian kernel (2). The
three examples differ in the value chosen for ? in the kernel - a large value (0.224 times the
diameter of the hemisphere) in the left figure and a small value (0.062 times the diameter
of the hemisphere) in the middle and right figure. In the right figure also non-zero slack
variables (outliers) were allowed. Note that that the outliers in the right figure correspond
to a sharp feature (non-smoothness) in the original surface.
equal 1 for most examples xi contained in the training set,2 while the regularization term
kwk will still be small. For an illustration, see Figure 1. The trade-off between these two
goals is controlled by a parameter ?.
One can show that the solution takes the form
f (x) = sgn
X
!
?i k(xi , x) ? ? ,
i
where the ?i are computed by solving the dual problem,
1X
?i ?j k(xi , xj )
minimize
m
??R
2 ij
subject to
0 ? ?i ?
X
1
and
?i = 1.
?m
i
(6)
(7)
(8)
Note that according to (8), the training examples contribute with nonnegative weights ? i ?
0 to the solution (6). One can show that asymptotically, a fraction ? of all training examples
will have strictly positive weights, and the rest will be zero (the ??-property?).
In our application we are not primarily interested in a decision function itself but in the
boundaries of the regions in input space defined by the decision function. That is, we are
interested in f ?1 (0), where f is the kernel expansion (3) and the points x1 , . . . , xm ? X
are sampled from some unknown hypersurface X ? Rd . We want to consider f ?1 (0) as
a model for X . In the following we focus on the case d = 3. If we assume that the x i are
sampled without noise from X ? which for example is a reasonable assumption for data
obtained with a state of the art 3d laser scanning device ? we should set the slack variables
in (4) and (5) to zero. In the dual problem this results in removing the upper constraints
on the ?i in (8). Note that sample points with non-zero slack variable cannot be contained
in f ?1 (0). But also sample points whose image in feature space lies above the optimal
hyperplane are not contained in f ?1 (0) (see Figure 1) ? we will address this in the next
section. It turns out that it is useful in practice to allow non-zero slack variables, because
they prevent f ?1 (0) from decomposing into many connected components (see Figure 2 for
an illustration).
In our experience, one can ensure that the images of all sample points in feature space lie
close to (or on) the optimal hyperplane can be achieved by choosing ? in the Gaussian
2
We use the convention that sgn (z) equals 1 for z ? 0 and ?1 otherwise.
Figure 3: Two parallel hyperplanes hw, ?(x)i = ? +
? (?) enclosing all but two
of the points. The outlier
?(x(?) ) is associated with
a slack variable ? (?) , which
is penalized in the objective
function (9).
? (x *)o
o
.
?*/ ||w||
o
o
o
(?+?* )/||w||
(?+?)/ ||w||
w
? (x) o
o
?/ ||w||
o
o
kernel (2) such that the Gaussians in the kernel expansion (3) are highly localized. However, highly localized Gaussians are not well suited for interpolation ? the implicit surface
decomposes into several components. Allowing outliers mitigates the situation to a certain
extent. Another way to deal with the problem is to further restrict the optimal region in
feature space. In the following we will pursue the latter approach.
3
Slab SVMs
A richer class of solutions, where some of the weights can be negative, is obtained if we
change the geometric setup. In this case, we estimate a region which is a slab in the RKHS,
i.e., the area enclosed between two parallel hyperplanes (see Figure 3).
To this end, we consider the following modified program:3
minimize
w?H,? (?) ?Rm ,??R
subject to
and
1 X
1
kwk2 +
(?i + ?i? ) ? ?
2
?m i
(9)
? ? ?i ? hw, ?(xi )i ? ? ? ? ? + ?i?
(10)
(?)
?i
(11)
? 0.
Here, ? (?) are fixed parameters. Strictly speaking, one of them is redundant: one can show
that if we subtract some offset from both, then we obtain the same overall solution, with ?
changed by the same offset. Hence, we can generally set one of them to zero, say, ? = 0.
Below we summarize some relationships of this convex quadratic optimization problem to
known SV methods:
1. For ? = 0 and ? ? = ? (i.e., no upper constraint), we recover the single-class SVM
(4)?(5).
2. If we drop ? from the objective function and set ? = ??, ? ? = ? (for some fixed ? ? 0),
we obtain the ?-insensitive support vector regression algorithm [11], for a data set where
all output values y1 , . . . , ym are zero. Note that in this case, the solution is trivial, w = 0.
This shows that the ? in our objective function plays an important role.
P
3. For ? = ? ? = 0, the term i (?i + ?i? ) measures the distance of the point ?(xi ) from the
hyperplane hw, ?(xi )i ? ? = 0 (up to a scaling of kwk). If ? tends to zero, this term will
dominate the objective function. Hence, in this case, the solution will be a hyperplane that
approximates the data well in the sense that the points lie close to it in the RKHS norm.
3
Here and below, the superscript (?) simultaneously denotes the variables with and without asterisk, e.g., ? (?) is a shorthand for ? and ? ? .
From the following constraints and Lagrange multipliers
?i ? ? + hw, ?(xi )i ? ? ? 0,
?i ? 0
?i? + ? ? + ? ? hw, ?(xi )i ? 0,
?i? ? 0
(?)
?i
? 0,
(?)
?i
?0
(12)
(13)
(14)
4
we derive the Lagrangian dual optimization problem of (9) - (11):
X
X
1X
?i + ? ?
?i?
(?i ? ?i? )(?j ? ?j? )k(xi , xj ) ? ?
minimize
m
??R
2 ij
i
i
subject to
and
1
(?)
0 ? ?i ?
?m
X
(?i ? ?i? ) = 1,
(15)
(16)
(17)
i
Note that for ? = ? ? , we can simplify the optimization problem using the transformation
?new = ? ? ?? . For ? = ? ? = 0, we thus obtain the single-class SVM (7) with the
1
1
modified box constraint ? ?m
? ?inew ? ?m
.
The dual problem can be solved using standard quadratic programming packages. The
offset ? can be computed from the value of the corresponding variable in the double dual, or
using the Karush-Kuhn-Tucker (KKT) conditions, just as in other support vector methods.
Once this is done, we can evaluate for each test point x whether it satisfies ? ? hw, ?(x)i?
? ? ? ? . In other words, we have an implicit description of the region in input space that
corresponds to the region in between the two hyperplanes in the RKHS. For ? = ? ? , this
is a single hyperplane, corresponding to a hypersurface in input space.5 To compute this
surface we use the kernel expansion
X
hw, ?(x)i =
(?i ? ?i? )k(xi , x).
(18)
i
Support Vectors and Outliers In our discussion of single class SVMs for surface modeling we already mentioned that we aim for many support vectors (as we want most training
points to lie on the surface) and that outliers might represent features like certain singularities in the original hypersurface.
Here we analyze how the parameter ? influences the SVs and outliers. To this end, we
introduce the following shorthands for the sets of SV and outlier indices:
SV := {i | hw, ?(xi )i ? ? ? ? ? 0}
(19)
SV ? := {i | hw, ?(xi )i ? ? ? ? ? ? 0}
(20)
(?)
OL(?) := {i | ?i > 0}
(21)
It is clear from the primal optimization problem that for all i, ?i > 0 implies hw, ?(xi )i ?
? ? ? < 0 (and likewise, ?i? > 0 implies hw, ?(xi )i ? ? ? ? ? > 0), hence OL(?) ? SV (?) .
The difference of the SV and OL sets are those points that lie precisely on the boundaries
of the constraints.6 Below, |A| denotes the cardinality of the set A.
4
Note that due to (17), the dual solution is invariant with respect to the transformation ? (?) ?
? (?) + const. ? such a transformation only adds a constant to the objective function, leaving the
solution unaffected.
5
subject to suitable conditions on k
6
The present usage differs slightly from the standard definition of SVs, which are usually those
(?)
that satisfy ?i > 0. In our definition, SVs are those points where the constraints are active. How(?)
ever, the difference is marginal: (i) It follows from the KKT conditions that ? i > 0 implies that
the corresponding constraint is active. (ii) while it can happen in theory that a constraint is active and
(?)
nevertheless the corresponding ?i is zero, this almost never occurs in practice.
Proposition 1 The solution of (9)?(11) satisfies
|OL| |SV ? |
|SV | |OL? |
?
???
?
.
m
m
m
m
(22)
The proof is analogous to the one of the ??-property? for standard SVMs, cf. [8]. Due to
lack of space, we skip it, and instead merely add the following observations:
1. The above statements are not symmetric with respect to exchanging the quantities with
asterisks and their counterparts without asterisk. This is due to the sign of ? in the primal
objective function. If we used +? rather than ??, we would obtain almost the same dual,
the only difference being that the constraint (17) would have a ??1? on the right hand
side. In this case, the role of the quantities with and without asterisks would be reversed in
Proposition 1.
2. The ?-property of single class SVMs is obtained as the special case where OL ? =
SV ? = ?.
3. Essentially, if we require that the distribution has a density w.r.t. the Lebesgue measure,
and that k is analytic and non-constant (cf. [8, 9]), it can be shown that asymptotically, the
two inequalities in the proposition become equalities with probability 1.
Implementation On larger problems, solving the dual with standard QP solvers becomes
too expensive (scaling with m3 ). For this case, we can use decomposition methods. The
adaptation of known decomposition methods to the present case is straightforward, noticing
that the dual of the standard ?-SV regression algorithm [11] becomes almost identical to the
present dual if we set ? = (? ? ? ?)/2 and yi = ?(? ? + ?)/2 for all i. The only difference
is that in our case, there is a ?1? in (17), whereas in the SVR case, we would have a ?0?.
As a consequence, we have to change the initialization of the optimization algorithm to
ensure that we start with a feasible solution. As an optimizer, we used a modified version
of libSVM [2].
Experimental Results In all our experiments we used a Gaussian kernel (2). To render
the implicit surfaces, i.e., the zero-set f ?1 (0), we generated a triangle mesh that approximates it. To compute the mesh we used an adaptation of the marching cubes algorithm [5]
which is a standard technique to transform an implicitly given surfaces into a mesh. The
most costly operations in the marching cubes algorithm are evaluations of the kernel expansion (18). To reduce the number of these evaluations we implemented a surface following
technique that exploits the fact that we know quite some sample points on the surface,
namely the support vectors.7 Some results can be seen in Figure 4.
Our experiments indicate a nice geometric interpretation of negative coefficients ? i ? ?i? .
It seems that negative coefficients correspond to concavities in the original model. The
coefficients seem well suited to extract shape features from the sample point set, e.g., the
detection of singularities like sharp edges or feature lines ? which is an important topic in
computer graphics [7].
We also tried a multi-scale approach. In this approach at first a rough model is computed
from ten percent of the sample points using a slab SVM. For the remaining 90% of the
sample points we compute the residual values, i.e., we evaluate the kernel expansion (18)
at the sample points. Finally we use support vector regression (SVR) and the residual
values to derive a new kernel expansion (using a smaller kernel width) whose zero set we
use as our surface model. An example how this approach works can be seen in Figure 5.
7
In the experiments, both the SVM optimization and the marching cubes rendering took up to
about 2 hours.
Figure 4: First row: Computing a model of the Stanford bunny (35947 points) and of a golf
club (16864 points) with the slab SVM. The close up of the ears and nose of the bunny
shows the sample points colored according to the coefficients ?i ? ?i? . Dark gray points
have negative coefficients and light gray points positive ones. In the right figure we show
the bottom of the golf club model. The model on the left of this figure was computed with
a different method [4]. Note that with this method fine details like the figure three become
visible. Such details get leveled out by the limited resolution of the marching cubes method.
However the information about these details is preserved and detected in the SVM solution,
as can be seen from the color coding. Second row: In the left and in the middle figure we
show the results of the slab SVM method on the screwdriver model (27152 points) and
the dinosaur model (13990 points), respectively. In the right figure a color coding of the
coefficients for the rockerarm data set (40177 points) is shown. Note that we can extract
sharp features from this data set by filtering the coefficients according to some threshold.
Figure 5: First row: The multi-scale approach applied to a knot data set (10000 points).
The blobby support surface (left figure) was computed from 1000 randomly chosen sample
points with the slab SVM. In the middle we show a color coding of the residual values of
all sample points (cf. http://books.nips.cc for color images). In the right figure we show the
surface that we get after applying support vector regression using the residual values.
4
Discussion and Outlook
An approximate description of the data as the zero set of a function can be useful as a compact representation of the data. It could potentially also be employed in other tasks where
models of the data are useful, such as denoising and image super-resolution. We therefore
consider it worthwhile to explore the algorithmic aspects of implicit surface estimation in
more depth, including the study of regression based approaches.
Some acquisition devices do not only provide us with points from a surface embedded
in R3 , but also with the normals at these points. Using methods similar to the ones in
[3], it should be possible to integrate such additional information into our approach. We
expect that it will improve the quality of the computed models in the sense that even more
geometric details are preserved.
A feature of our approach is that its complexity depends only marginally on the dimension
of the input space (in our examples this was three). Thus the approach should work also
well for hypersurfaces in higher dimensional input spaces. From an applications point of
view hypersurfaces might not be as interesting as manifolds of higher co-dimension. It
would be interesting to see if our approach can be generalized to handle also this situation.
Acknowledgment We thank Chih-Jen Lin for help with libSVM. The bunny data were
taken from the Stanford 3d model repository. The screwdriver, dinosaur and rockerarm
data were taken from the homepage of Cyberware Inc. Thanks to Koby Crammer, Florian
Steinke, and Christian Walder for useful discussion.
References
[1] J. Carr, R. Beatson, J. Cherrie, T. Mitchell, W. Fright, B. McCallum, and T. Evans.
Reconstruction and representation of 3D objects with radial basis functions. In Proc.
28th Ann. Conf. Computer Graphics and Interactive Techniques, pages 67?76. 2001.
[2] C.-C. Chang and C.-J. Lin. LIBSVM: a library for support vector machines, 2001.
Software available at http://www.csie.ntu.edu.tw/?cjlin/libsvm.
[3] O. Chapelle and B. Sch?olkopf. Incorporating invariances in nonlinear SVMs. In T.G.
Dietterich, S. Becker, and Z. Ghahramani, editors, Advances in Neural Information
Processing Systems 14, Cambridge, MA, 2002. MIT Press.
[4] J. Giesen and M. John. Surface reconstruction based on a dynamical system. Computer Graphics Forum, 21(3):363?371, 2002.
[5] T. Lewiner, H. Lopes, A. Wilson, and G. Tavares. Efficient implementation of marching cubes cases with topological guarantee. Journal of Graphics Tools, 8:1?15, 2003.
[6] S. Osher and N. Paragios. Geometric Level Set Methods. Springer, New York, 2003.
[7] M. Pauly, R. Keiser, and M. Gross. Multi-scale feature extraction on point-sampled
surfaces. Computer Graphics Forum, 22(3):281?289, 2003.
[8] B. Sch?olkopf, J. Platt, J. Shawe-Taylor, A. J. Smola, and R. C. Williamson. Estimating
the support of a high-dimensional distribution. Neural Computation, 13:1443?1471,
2001.
[9] I. Steinwart. Sparseness of support vector machines?some asymptotically sharp
bounds. In S. Thrun, L. Saul, and B. Sch?olkopf, editors, Advances in Neural Information Processing Systems 16. MIT Press, Cambridge, MA, 2004.
[10] D. M. J. Tax and R. P. W. Duin. Support vector data description. Machine Learning,
54:45?66, 2004.
[11] V. N. Vapnik. The Nature of Statistical Learning Theory. Springer Verlag, New York,
1995.
| 2724 |@word repository:1 briefly:1 version:1 middle:3 norm:2 seems:1 tried:1 decomposition:2 outlook:1 solid:2 contains:1 rkhs:5 john:1 mesh:4 visible:1 happen:1 evans:1 shape:2 analytic:1 christian:1 drop:1 half:1 device:3 mccallum:1 colored:1 contribute:1 club:2 hyperplanes:4 unbounded:1 differential:1 become:2 shorthand:2 inside:2 introduce:1 x0:3 mpg:1 multi:3 ol:6 cardinality:1 solver:1 becomes:2 project:1 estimating:1 bounded:1 homepage:1 pursue:1 transformation:4 guarantee:1 interactive:1 k2:1 rm:2 platt:1 planck:1 positive:4 tends:1 consequence:1 interpolation:1 approximately:2 might:2 initialization:1 co:1 limited:1 range:1 acknowledgment:1 testing:1 practice:2 definite:2 differs:1 swiss:1 area:1 intersect:1 eth:1 word:1 radial:1 get:2 svr:2 cannot:1 unlabeled:1 close:3 influence:1 applying:1 www:1 map:3 lagrangian:1 straightforward:1 normed:1 convex:1 resolution:2 dominate:1 handle:1 notion:1 analogous:1 suppose:1 today:1 play:1 programming:1 origin:4 expensive:1 bottom:1 role:2 csie:1 solved:1 capture:1 region:6 schoelkopf:1 connected:1 trade:1 mentioned:1 gross:1 transforming:1 complexity:1 solving:3 basis:1 triangle:2 easily:1 derivation:1 laser:2 describe:2 query:2 detected:1 outside:2 choosing:1 whose:2 richer:1 larger:1 valued:1 solve:1 say:1 quite:1 otherwise:1 stanford:2 unseen:1 transform:2 itself:1 superscript:1 advantage:3 took:1 reconstruction:2 adaptation:2 beatson:1 tu:1 tax:1 intuitive:1 description:3 olkopf:4 double:1 object:1 help:1 depending:1 derive:2 measured:1 ij:2 implemented:1 skip:1 implies:4 indicate:1 convention:1 differ:1 switzerland:1 greek:1 kuhn:1 sgn:3 require:1 inew:1 karush:1 ntu:1 proposition:3 biological:1 singularity:2 strictly:2 scanner:1 sufficiently:2 normal:1 exp:1 mapping:1 algorithmic:1 slab:6 optimizer:1 smallest:1 giesen:3 estimation:1 proc:1 tool:1 rough:1 mit:2 gaussian:4 aim:1 modified:3 rather:1 super:1 wilson:1 focus:1 joachim:1 sense:2 interested:2 germany:1 overall:1 dual:10 art:1 special:1 marginal:1 equal:2 construct:1 once:1 never:1 cube:5 sampling:2 extraction:1 identical:1 blobby:1 koby:1 simplify:1 primarily:1 randomly:1 simultaneously:1 national:1 lebesgue:1 detection:1 highly:2 evaluation:3 light:1 primal:2 edge:1 screwdriver:2 necessary:1 experience:1 euclidean:1 loosely:1 taylor:1 modeling:2 exchanging:1 too:1 graphic:5 scanning:1 sv:10 thanks:1 density:1 off:1 ym:1 connectivity:1 ear:1 containing:1 conf:1 book:1 toy:1 de:1 sizeable:1 bold:1 coding:3 coefficient:7 inc:1 satisfy:1 depends:1 leveled:1 view:1 kwk:5 analyze:1 start:1 recover:1 parallel:2 simon:1 minimize:4 likewise:1 correspond:2 knot:1 marginally:1 cc:1 cybernetics:1 unaffected:1 definition:2 acquisition:2 tucker:1 associated:5 proof:1 boil:1 sampled:4 popular:2 mitchell:1 color:4 hilbert:4 higher:3 day:1 done:1 box:1 just:1 implicit:9 smola:1 hand:1 steinwart:1 nonlinear:1 lack:1 quality:1 gray:2 resemblance:1 usage:1 dietterich:1 contain:1 true:1 multiplier:1 counterpart:1 regularization:1 hence:3 equality:1 symmetric:1 deal:1 width:1 generalized:1 carr:1 percent:1 image:4 recently:1 qp:1 insensitive:1 interpretation:1 approximates:2 kwk2:2 cambridge:2 smoothness:1 rd:2 shawe:1 chapelle:1 similarity:1 surface:21 add:3 curvature:1 recent:1 inf:2 hemisphere:2 certain:2 verlag:1 inequality:1 yi:1 seen:3 pauly:1 additional:1 florian:1 employed:1 redundant:1 ii:1 smooth:2 lin:2 controlled:1 regression:5 tavares:1 vision:1 essentially:1 kernel:19 represent:1 achieved:1 bunny:3 whereas:1 want:3 fine:1 preserved:2 leaving:1 sch:4 rest:1 subject:5 seem:1 rendering:1 xj:2 restrict:1 reduce:1 idea:2 golf:2 whether:2 becker:1 render:1 bingen:1 speaking:2 york:2 svs:3 useful:4 generally:1 clear:1 amount:1 dark:1 ten:1 svms:7 diameter:2 http:2 fright:1 sign:2 popularity:1 nevertheless:1 threshold:1 prevent:1 libsvm:4 cherrie:1 asymptotically:3 merely:1 fraction:3 run:1 package:1 noticing:1 lope:1 family:1 reasonable:2 almost:3 chih:1 separation:1 decision:3 scaling:2 bound:1 quadratic:3 topological:1 nonnegative:1 duin:1 constraint:9 precisely:1 software:1 aspect:1 department:1 according:4 smaller:1 slightly:1 character:1 tw:1 osher:1 outlier:10 invariant:1 taken:2 previously:1 slack:7 turn:1 r3:1 cjlin:1 know:1 nose:1 end:2 available:1 decomposing:1 gaussians:2 operation:1 worthwhile:1 original:3 denotes:2 remaining:1 ensure:2 cf:3 const:1 exploit:2 quantile:1 especially:1 ghahramani:1 forum:2 objective:8 already:1 quantity:3 occurs:1 costly:1 distance:3 separate:2 mapped:1 reversed:1 thank:1 thrun:1 topic:1 manifold:2 extent:1 tuebingen:1 trivial:1 modeled:1 index:2 illustration:2 relationship:1 setup:1 statement:1 potentially:1 negative:4 enclosing:1 implementation:2 unknown:2 allowing:1 upper:2 observation:1 finite:2 walder:1 situation:2 ever:1 y1:1 reproducing:2 sharp:4 introduced:1 namely:1 specified:2 hour:1 nip:1 address:1 below:4 usually:1 xm:3 dynamical:1 summarize:1 program:2 max:1 including:1 suitable:1 residual:4 smallness:2 improve:1 library:1 extract:2 nice:1 geometric:5 determining:2 embedded:1 expect:2 bear:1 interesting:3 filtering:1 enclosed:1 localized:2 asterisk:4 foundation:1 integrate:1 editor:2 row:3 dinosaur:2 penalized:3 changed:1 supported:1 side:2 allow:3 institute:1 steinke:1 saul:1 face:1 boundary:3 default:1 depth:1 dimension:2 rich:1 computes:1 concavity:1 coincide:1 hypersurface:5 approximate:1 compact:1 implicitly:1 bernhard:2 kkt:2 active:3 xi:16 continuous:1 decomposes:1 nature:1 expansion:7 williamson:1 domain:1 main:1 noise:1 allowed:1 x1:3 quantiles:1 paragios:1 lie:6 hw:14 down:1 removing:1 jen:1 zu:1 mitigates:1 offset:3 svm:10 incorporating:1 vapnik:1 gained:1 sparseness:1 margin:2 kx:1 marching:5 subtract:1 suited:2 depicted:1 explore:1 lagrange:1 expressed:1 contained:3 partially:1 chang:1 springer:2 ch:2 corresponds:2 hypersurfaces:2 satisfies:2 ma:2 goal:2 ann:1 feasible:1 change:2 hyperplane:8 denoising:1 called:1 invariance:1 experimental:1 attempted:1 m3:1 support:12 latter:2 crammer:1 ethz:2 evaluate:2 |
1,899 | 2,725 | A Machine Learning Approach
to Conjoint Analysis
Olivier Chapelle, Za??d Harchaoui
Max Planck Institute for Biological Cybernetics
Spemannstr. 38 - 72076 T?ubingen - Germany
{olivier.chapelle,zaid.harchaoui}@tuebingen.mpg.de
Abstract
Choice-based conjoint analysis builds models of consumer preferences
over products with answers gathered in questionnaires. Our main goal is
to bring tools from the machine learning community to solve this problem more efficiently. Thus, we propose two algorithms to quickly and
accurately estimate consumer preferences.
1
Introduction
Conjoint analysis (also called trade-off analysis) is one of the most popular marketing research technique used to determine which features a new product should have, by conjointly
measuring consumers trade-offs between discretized1 attributes. In this paper, we will focus on the choice-based conjoint analysis (CBC) framework [11] since it is both widely
used and realistic: at each question in the survey, the consumer is asked to choose one
product from several.
The preferences of a consumer are modeled via a utility function representing how much a
consumer likes a given product. The utility u(x) of a product x is assumed to be the sum
of the partial utilities (or partworths) for each attribute, i.e. linear: u(x) = w ? x. However,
instead of observing pairs (xl , yl ), the training samples are of the form ({x1k , . . . , xpk }, yk )
indicating that among the p products {x1k , . . . , xpk }, the ykth was preferred. Without noise,
this is expressed mathematically by u(xykk ) ? u(xbk ), ? b 6= yk .
Let us settle down the general framework of a regular conjoint analysis survey. We have a
population of n consumers available for the survey. The survey consists of a questionnaire
of q questions for each consumer, each asking to choose one product from a basket of
p. Each product profilePis described through a attributes with l1 , ..., la levels each, via a
a
vector of length m = s=1 ls , with 1 at positions of levels taken by each attribute and 0
elsewhere.
Marketing researchers are interested in estimating individual partworths in order to perform for instance a segmentation of the population afterwards. But traditional conjoint
estimation techniques are not reliable for this task since the number of parameters m to be
estimated is usually larger than the number of answers q available for each consumer. They
estimate instead the partworths on the whole population (aggregated partworths). Here we
1
e.g. if the discretized attribute is weight, the levels would be light/heavy.
aim to investigate this issue, for which machine learning can provide efficient tools. We
also address adaptive questionnaire design with active learning heuristics.
2
Hierarchical Bayes Analysis
The main idea of HB2 is to estimate the individual utility functions under the constraint that
their variance should not be too small. By doing so, the estimation problem is not ill-posed
and the lack of information for a consumer can be completed by the other ones.
2.1
Probabilistic model
In this section, we follow [11] for the description of the HB model and its implementation.
This method aims at estimating the individual linear utility functions ui (x) = wi ? x, for
1 ? i ? n. The probabilistic model is the following:
1. The individual partworths wi are drawn from a Gaussian distribution with mean ?
(representing the aggregated partworths) and covariance ? (encoding population?s
heterogeneity),
2. The covariance matrix ? has an invert Wishart prior, and ? has an (improper) flat
prior.
3. Given a set of products (x1 , . . . xp ), the probability that the consumer i chooses
the product x? is given by
exp(wi ? x? )
P (x? |wi ) = Pp
.
b=1 exp(wi ? xb )
2.2
(1)
Model estimation
We describe now the standard way of estimating ?, w ? (w1 , . . . , wn ) and ? based on
Gibbs sampling and then propose a much faster algorithm that approximates the maximum
a posteriori (MAP) solution.
Gibbs sampling As far as we know, all implementations of HB rely on a variant of the
Gibbs sampling [11]. During one iteration, each of the three sets of variables (?, w and ?)
is drawn in turn from its posterior distribution the two others being fixed. Sampling for ?
and ? is straightforward, whereas sampling from P (w|?, ?, Y ) ? P (Y |w). P (w|?, ?)
is achieved with the Metropolis-Hastings algorithm.
When convergence is reached, the sampling goes on and finally outputs the empirical expectation of ?, w and ?. Although the results of this sampling-based implementation of
HB3 are impressive, practitioners complain about its computational burden.
Approximate MAP solution So far HB implementations make predictions by evaluating
(1) at the empirical mean of the samples, in contrast with the standard bayesian approach,
which would average the rhs of (1) over the different samples, given samples w from the
posterior. In order to alleviate the computational issues associated with Gibbs sampling,
we suggest to consider the maximum of the posterior distribution (maximum a posteriori,
MAP) rather than its mean.
2
Technical papers of Sawtooth software [11], the world leading company for conjoint analysis
softwares, provide very useful and extensive references.
3
that we will call HB-Sampled or HB-S in the rest of the paper.
To find ?, w and ? which maximize P (?, w, ?|Y ), let us use Bayes? rule,
P (?, w, ?|Y ) ? P (Y |?, w, ?) ? P (w|?, ?) ? P (?|?) ? P (?)
? P (Y |w) ? P (w|?, ?) ? P (?)
(2)
w
Maximizing (2) with respect to ? yields ?MAP = I+C
n+d , with Cw being the ?covariance?
P
matrix of the wi centered at ?: Cw = (wi ? ?)(wi ? ?)> . Putting back this value in
(2), we get
? log P (?, w, ?MAP |Y ) = ? log P (Y |w) + log |I + Cw (?)| + C,
(3)
where C is an irrelevant constant. Using the model (1), the first term in the rhs of (3) is
convex in w, but P
not the second term. For this reason, we propose to change log |I + Cw |
by trace(Cw ) =
||wi ? ?||2 (this would be a valid approximation if trace(Cw ) 1).
With this new prior on w, the rhs of (3) becomes
W (?, w) =
n
X
? log P (Yi |wi ) + ||wi ? ?||2 .
(4)
i=1
As in equation (3), this objective function is minimized with respect to ? when ? is equal to
the empirical mean of the wi . We thus suggest the following iterative scheme to minimize
the convex functional (4):
1. For a given ?, minimize (4) with respect to each of the wi independently.
2. For a given w, set ? to the empirical mean4 of the w.
Thanks to the convexity, this optimization problem can be solved very efficiently. A Newton approach in step 1, as well as in step 2 to speed-up the global convergence to a fixed
point ?, has been implemented. Only couple of steps in both cases are necessary to reach
convergence.
Remark The approximation from equation (3) to (4) might be too crude. After all it boils
down to setting ? to the identity matrix. One might instead consider ? as an hyperparameter and optimize it by maximizing the marginalized likelihood [14].
3
Conjoint Analysis with Support Vector Machines
Similarly to what has recently been proposed in [3], we are now investigating the use of
Support Vector Machines (SVM) [1, 12] to solve the conjoint estimation problem.
3.1
Soft margin formulation of conjoint estimation
Let us recall the learning problem. At the k-th question, the consumer chooses the ykth
product from the basket {x1k , . . . , xpk }: w ? xykk ? w ? xbk , ? b 6= yk . Our goal is to estimate
the individual partworths w, with the individual utility function now being u(x) = w ? x.
With a reordering of the products, we can actually suppose that yk = 1. Then the above
inequalities can be rewritten as a set of p ? 1 constraints:
w ? (x1k ? xbk ) ? 0, 2 ? b ? p.
(5)
Eq. (5) shows that the conjoint estimation problem can be cast as a classification problem
in the product-profiles differences space. From this point of view, it seems quite natural to
use state-of-the-art classifiers such as SVMs for this purpose.
4
which is consistent with the L2 -loss measuring deviations of wi -s from ?.
More specifically, we propose to train a L2 -soft margin classifier (see also [3] for a similar
approach) with only positive examples and with a hyperplane passing through the origin
(no bias), modelling the noise in the answers with slack variables ?kb :
Pq Pp
2
Minimize w2 + C k=1 b=2 ?kb
1
b
subject to w ? (xk ? xk ) ? 1 ? ?kb .
3.2
Estimation of individual utilities
It was proposed in [3] to train one SVM per consumer to get wi and to compute
Pn the
individual partworths by regularizing with the aggregated partworths w = n1 i=1 wi :
wi? = wi2+w .
Instead, to estimate the individual utility partworths wi , we suggest the following optimization problem (the set Qi contains the indices j such that the consumer i was asked to
choose between products x1k , . . . , xpk ) :
(
P
Pp
P
Pp
?
2
2
Minimize wi2 + qCi k?Qi b=2 ?kb
+ P C qj k?Q
/ i
b=2 ?kb
j6=i
subject to wi ? (x1k ? xbk ) ? 1 ? ?kb , ?k, ?b ? 2 .
C
Here the ratio C
? determines the trade-off between the individual scale and the aggregated
C
5
one. For C? = 1, the population is modeled as if it were homogeneous, i.e. all partworths
wi are equal. For C
? 1, the individual partworths are computed independently, without
C
taking into account aggregated partworths.
4
Related work
Ordinal regression Very recently [2] explores the so-called ordinal regression task for
ranking, and derive two techniques for hyperparameters learning and model selection in
a hierarchical bayesian framework, Laplace approximation and Expectation Propagation
respectively. Ordinal regression is similar yet distinct from conjoint estimation since training data are supposed to be rankings or ratings in contrast with conjoint estimation where
training data are choice-based. See [4] for more extensive bibliography.
Large margin classifiers Casting the preference problem in a classification framework,
leading to learning by convex optimization, was known for a long time in the psychometrics community. [5] pioneered the use of large margin classifiers for ranking tasks.
[3] introduced the kernel methods machinery for conjoint analysis on the individual scale.
Very recently [10] proposes an alternate method for dealing with heterogeneity in conjoint
analysis, which boils down to a very similar optimization to our HB-MAP approximation
objective function, but with large margin regularization and with minimum deviation from
the aggregated partworths.
Collaborative filtering Collaborative filtering exploits similarity between ratings across
a population. The goal is to predict a person?s rating on new products given the person?s
past ratings on similar products and the ratings of other people on all the products. Again
collaborative is designed for overlapping training samples for each consumer, and usually
rating/ranking training data, whereas conjoint estimation usually deals with different questionnaires for each consumer and choice-based training data.
5
? In this way, directions for which the xj , j ? Qi contain information are estimated
C ? C
accurately, whereas the others directions are estimated thanks to the answers of the other consumers.
5
Experiments
Artificial experiments We tested our algorithms on the same benchmarking artificial experimental setup used in [3, 16]. The simulated product profiles consist of 4 attributes, each
of them being discretized through 4 levels. A random design was used for the questionnaire. For each question, the consumer was asked to choose one product from a basket of 4.
A population of 100 consumers was simulated, each of them having to answer 4 questions.
Finally, the results presented below are averaged over 5 trials.
The 100 true consumer partworths were generated from a Gaussian distribution with mean
(??, ??/3, ?/3, ?) (for each attribute) and with a diagonal covariance matrix ? 2 I. Each
answer is a choice from the basket of products, sampled from the discrete logit-type distribution (1). Hence when ? (called the magnitude6 ) is large, the consumer will choose with
high probability the product with the highest utility, whereas when ? is small, the answers
will be less reliable. The ratio ? 2 /? controls the heterogeneity7 of the population.
Finally, as in [3], the performances are computed using the mean of the L2 distances between the true and estimated individual partworths (also called RMSE). Beforehand the
partworths are translated such that the mean on each attribute is 0 and normalized to 1.
Real experiments We tested our algorithms on disguised industrial datasets kindly provided by Sawtooth Software Inc., the world leading company in conjoint analysis softwares.
11 one-choice-based8 conjoint surveys datasets9 were used for real experiments below. The
number of attributes ranged from 3 to 6 (hence total number of levels from 13 to 28), the
size of the baskets, to pick one product from at each question, ranged from 2 to 5, and
the number of questions ranged from 6 to 15. The numbers of respondents ranged roughly
from 50 to 1200. Since here we did not address the issue of no choice options in question
answering, we removed10 questions where customers refused to choose a product from the
basket and chose the no-choice-option as an answer11 .
Finally, as in [16], the performances are computed using the hit rate, i.e. the misprediction
rate of the preferred product.
5.1
Analysis of HB-MAP
We compare in this section our implementation of the HB method described in Section 2,
that we call HB-MAP, to HB-S, the standard HB implementation.
The average training time for HB-S was 19 minutes (with 12000 iterations as suggested in
[11]), whereas our implementation based on the approximation of the MAP solution took
in average only 1.8 seconds. So our primary goal, i.e. to alleviate the sampling phase
complexity, was achieved since we got a speed-up factor of the order of 1000.
The accuracy does not seem to be significantly weakened by this new implementation.
Indeed, as shown in both Table 1 and Table 2, the performances achieved by HB-MAP
were surprisingly often as good as HB-S?s, and sometimes even a bit better. This might be
6
as in [3], we tested High Magnitude (? = 3) and Low Magnitude (? = 0.5).
It was either set to ? 2 = 3? or ? 2 = 0.5?, respectively High and Low Heterogeneity cases.
8
We limited ourselves to datasets in which respondents were asked to choose 1 product among a
basket at each question.
9
see [4] for more details on the numerical features of the datasets.
10
One could use EM-based methods to deal with such missing training choice data.
11
When this procedure boiled down to unreasonable number of questions for hold-out evaluation
of our algorithms, we simply removed the corresponding individuals.
7
explained by the fact that assuming that the covariance matrix is quasi-diagonal is a reasonable approximation, and that the mode of the posterior distribution is actually roughly close
to the mean, for the real datasets considered. Additionally it is likely that HB-S may have
demanded much more iterations for convergence to systematically behave more accurately
than HB-MAP as one would have normally expected.
5.2
Analysis of SVMs
We now turn to the SVM approach presented in section 3.2 that we call Im.SV12 . We did
not use a non-linear kernel in our experiments. Hence it was possible to minimize (3.2)
directly in the primal, instead of using the dual formulation as done usually. This turned
out to be faster since the number of constraints was, for our problem, larger than the number
of variables. The resulting mean training time was 4.7 seconds. The so-called chapspan,
span estimate of leave-one-out prediction error [17], was used to select a suitable value of
C 13 , since it gave a quasi-convex estimation on the regularization path.
The performances of Im.SV in Table 2, compared to the HB methods and logistic regression
[3] are very satisfactory in case of artificial experiments. In real experiments, Im.SV gives
overall quite satisfactory results, but sometimes disappointing ones in Table 2. One reason
? were optimized once for the whole population. This
might be that hyperparameters (C, C)
may also be due to the lack of robustness14 of Im.SV to heterogeneity in the number of
training samples for each consumer.
Table 1: Average RMSE between estimated and true individual partworths
Mag
L
L
H
H
Het
L
H
L
H
HB-S
0.90
0.95
0.44
0.72
HB-MAP
0.83
0.91
0.40
0.68
Logistic
0.84
1.16
0.43
0.82
Im.SV
0.86
0.90
0.41
0.67
Table 2: Hit rate performances on real datasets.
Dat12
Dat22
Dat13
Dat23
Dat33
Dat43
Im.SV
0.16
0.15
Im.SV
0.37
0.34
0.35
0.35
HB-MAP
0.16
0.13
HB-MAP
0.24
0.33
0.28
0.31
HB-S
0.17
0.15
HB-S
0.25
0.33
0.24
0.28
Dat15
Dat25
Im.SV
0.52
0.58
HB-MAP
0.45
0.47
HB-S
0.48
0.51
Dat14
Dat24
Dat34
Im.SV
0.33
0.33
0.45
HB-MAP
0.36
0.36
0.40
HB-S
0.35
0.28
0.25
Legend of Tables 1 and 2 The first two columns indicate the Magnitude and the
Heterogeneity (High or Low). p in Datmp is the number of products respondents are
asked to choose one from at each question.
12
since individual choice data are Immersed in the rest of the population choice data, via the
optimization objective
13
? was irrelevant, and that only the ratio C/C
? matWe observed that the value of the constant C
tered.
14
Indeed the no-choice data cleaning step might have lead to a strong unbalance to which Im.SV
is maybe much more sensitive than HB-MAP or HB-S.
6
Active learning
Motivation Traditional experimental designs are built by minimizing the variance of an
estimator (e.g. orthogonal designs [6]). However, they are sub-optimal because they do
not take into account the previous answers of the consumer. Therefore adaptive conjoint
analysis was proposed [11, 16] for adaptively designing questionnaires.
The adaptive design concept is often called active learning in machine learning, as the
algorithm can actively select questions whose responses are likely to be informative. In the
SVM context, a common and intuitive strategy is to select, as the next point to be labeled,
the nearest one from the decision boundary (see for instance [15]).
Experiments We implemented this heuristic for conjoint analysis by selecting for each
question a set of products whose estimated utilities are as close as possible15 . To compare
the different designs, we used the same artificial simulations as in section 5, but with 16
questions per consumer in order to fairly compare to the orthogonal design.
Table 3: Comparison of the RMSE achieved by different designs.
Mag
L
L
H
H
Het
L
H
L
H
Random
0.66
0.62
0.31
0.49
Orthogonal
0.61
0.56
0.29
0.45
Adaptive
0.66
0.56
0.24
0.34
Results in Table 3 show that active learning produced an adaptive design which seems
efficient, especially in the case of high magnitude, i.e. when the answers are not noisy16 .
7
Discussion
We may need to capture correlations between attributes to model interaction effects among
them. The polynomial kernel K(u, v) = (u.v + 1)d seems particularly relevant for such a
task. HB methods kernelization can be done in the framework presented in [7]. For large
margin methods [10, 3] give a way to use the kernel trick in the space of product-profiles
differences. Prior knowledge of product-profile structure [3] may also be incorporated in
the estimation process by using virtual examples [12].
[9] approach would allow us to improve our approximate MAP solution by learning a variational approximation of a non-isotropic diagonal covariance matrix.
A fully bayesian HB setting, i.e. with a maximum likelihood type II17 (ML II) step, in
contrast of sampling from the posterior, is known in the statistics community as bayesian
multinomial logistic regression. [18] use Laplace approximation to compute integration
over hyperparameters for multi-class classification, while [8] develop a variational approximation of the posterior distribution.
New insights on learning gaussian process regression in a HB framework have just been
given in [13], where a method combining an EM algorithm and a generalized Nystr?om approximation of covariance matrix is proposed, and could be incorporated in the HB-MAP
approximation above.
15
Since the bottom-line goal of the conjoint analysis is not really to estimate the partworths but
to design the ?optimal? product, adaptive design can also be helpful by focusing on products which
have a high estimated utility.
16
Indeed noisy answers are neither informative nor reliable for selecting the next question.
17
aka evidence maximization or hyperparameters learning
8
Conclusion
Choice-based conjoint analysis seems to be a very promising application field for machine
learning techniques. Further research include fully bayesian HB methods, extensions to
non-linear models as well as more elaborate and realistic active learning schemes.
Acknowledgments
The authors are very grateful to J. Qui?nonero-Candela and C. Rasmussen for fruitful discussions, and O. Toubia for providing us with his HB implementation. Many thanks to
Sawtooth Software Inc. for providing us with real conjoint analysis datasets.
References
[1] B. E. Boser, I. M. Guyon, and V. N. Vapnik. A training algorithm for optimal margin
classifiers. In Proc. 5th Annu. Workshop on Comput. Learning Theory, 1992.
[2] W. Chu and Z. Ghahramani. Gaussian processes for ordinal regression. Technical
report, University College London, 2004.
[3] T. Evgeniou, C. Boussios, and G. Zacharia. Generalized robust conjoint estimation.
Marketing Science, 25, 2005.
[4] Z. Harchaoui. Statistical learning approaches to conjoint estimation. Technical report,
Max Planck Institute for Biological Cybernetics, to appear.
[5] R. Herbrich, T. Graepel, and K. Obermayer. Large margin rank boundaries for ordinal
regression. In Advances in Large Margin Classifiers. MIT Press, 2000.
[6] J. Huber and K. Zwerina. The importance of utility balance in efficient choice designs.
Journal of Marketing Research, 33, 1996.
[7] T. S. Jaakkola and D. Haussler. Probabilistic kernel regression models. In Artificial
Intelligence and Statistics, 1999.
[8] T. S. Jaakkola and M. I. Jordan. Bayesian logistic regression: a variational approach.
Statistics and Computing, 10:25?37, 2000.
[9] T. Jebara. Convex invariance learning. In Artificial Intelligence and Statistics, 2003.
[10] C. A. Micchelli and M. Pontil. Kernels for multi?task learning. In Advances in Neural
Information Processing Systems 17, 2005.
[11] Sawtooth Software.
Research paper series.
Available at
www.sawtoothsoftware.com/techpap.shtml#hbrel.
[12] B. Sch?olkopf and A. Smola. Learning with kernels. MIT Press, 2002.
[13] A. Schwaighofer, V. Tresp, and K. Yu. Hierarchical bayesian modelling with gaussian
processes. In Advances in Neural Information Processing Systems 17, 2005.
[14] M. Tipping. Bayesian inference: Principles and practice. In Advanced Lectures on
Machine Learning. Springer, 2004.
[15] S. Tong and D. Koller. Support vector machine active learning with applications to
text classification. In Journal of Machine Learning Research, volume 2, 2001.
[16] O. Toubia, J. R. Hauser, and D. I. Simester. Polyhedral methods for adaptive choicebased conjoint analysis. Journal of Marketing Research, 41(1):116?131, 2004.
[17] V. Vapnik and O. Chapelle. Bounds on error expectation for support vector machines.
Neural Computation, 12(9), 2000.
[18] C. K. I. Williams and D. Barber. Bayesian classification with gaussian processes.
IEEE Trans. Pattern Anal. Mach. Intell., 20, 1998.
| 2725 |@word trial:1 polynomial:1 seems:4 logit:1 simulation:1 covariance:7 pick:1 nystr:1 contains:1 series:1 selecting:2 mag:2 past:1 com:1 yet:1 chu:1 numerical:1 realistic:2 informative:2 zaid:1 designed:1 intelligence:2 xk:2 isotropic:1 preference:4 herbrich:1 consists:1 polyhedral:1 huber:1 expected:1 indeed:3 roughly:2 mpg:1 nor:1 multi:2 discretized:2 company:2 becomes:1 provided:1 estimating:3 misprediction:1 what:1 classifier:6 hit:2 control:1 normally:1 appear:1 planck:2 positive:1 encoding:1 mach:1 path:1 might:5 chose:1 weakened:1 limited:1 averaged:1 acknowledgment:1 practice:1 procedure:1 pontil:1 empirical:4 got:1 significantly:1 regular:1 suggest:3 get:2 close:2 selection:1 context:1 www:1 optimize:1 map:19 customer:1 missing:1 maximizing:2 fruitful:1 straightforward:1 go:1 williams:1 l:1 convex:5 survey:5 independently:2 rule:1 estimator:1 insight:1 haussler:1 his:1 population:10 laplace:2 suppose:1 pioneered:1 cleaning:1 olivier:2 homogeneous:1 designing:1 origin:1 trick:1 particularly:1 labeled:1 observed:1 bottom:1 solved:1 capture:1 improper:1 trade:3 highest:1 removed:1 yk:4 questionnaire:6 convexity:1 ui:1 complexity:1 asked:5 grateful:1 translated:1 train:2 distinct:1 describe:1 london:1 artificial:6 disguised:1 quite:2 heuristic:2 widely:1 solve:2 larger:2 posed:1 psychometrics:1 whose:2 statistic:4 noisy:1 took:1 propose:4 interaction:1 product:31 turned:1 relevant:1 combining:1 nonero:1 supposed:1 description:1 intuitive:1 olkopf:1 convergence:4 leave:1 derive:1 develop:1 nearest:1 eq:1 strong:1 implemented:2 indicate:1 direction:2 attribute:10 kb:6 centered:1 settle:1 virtual:1 really:1 alleviate:2 biological:2 mathematically:1 im:10 extension:1 hold:1 considered:1 exp:2 predict:1 purpose:1 estimation:14 proc:1 sensitive:1 tool:2 offs:1 mit:2 gaussian:6 aim:2 rather:1 pn:1 casting:1 jaakkola:2 shtml:1 focus:1 modelling:2 likelihood:2 rank:1 aka:1 contrast:3 industrial:1 posteriori:2 helpful:1 inference:1 koller:1 quasi:2 interested:1 germany:1 issue:3 among:3 ill:1 classification:5 dual:1 overall:1 proposes:1 art:1 integration:1 fairly:1 equal:2 once:1 field:1 having:1 evgeniou:1 sampling:10 yu:1 minimized:1 others:2 report:2 intell:1 individual:16 phase:1 ourselves:1 n1:1 investigate:1 evaluation:1 light:1 primal:1 xb:1 beforehand:1 partial:1 necessary:1 machinery:1 orthogonal:3 instance:2 column:1 soft:2 asking:1 measuring:2 maximization:1 deviation:2 too:2 hauser:1 answer:10 sv:9 chooses:2 adaptively:1 thanks:3 person:2 explores:1 probabilistic:3 off:2 yl:1 quickly:1 w1:1 again:1 choose:8 wishart:1 leading:3 actively:1 account:2 de:1 inc:2 ranking:4 toubia:2 view:1 candela:1 observing:1 doing:1 reached:1 bayes:2 option:2 rmse:3 collaborative:3 minimize:5 om:1 accuracy:1 variance:2 efficiently:2 gathered:1 yield:1 bayesian:9 boiled:1 accurately:3 produced:1 researcher:1 cybernetics:2 j6:1 za:1 reach:1 basket:7 pp:4 associated:1 boil:2 couple:1 sampled:2 tered:1 popular:1 recall:1 knowledge:1 segmentation:1 graepel:1 actually:2 back:1 focusing:1 tipping:1 follow:1 response:1 formulation:2 done:2 marketing:5 just:1 smola:1 correlation:1 hastings:1 overlapping:1 lack:2 propagation:1 sv12:1 mode:1 logistic:4 effect:1 contain:1 true:3 normalized:1 ranged:4 concept:1 regularization:2 hence:3 satisfactory:2 deal:2 during:1 generalized:2 l1:1 bring:1 regularizing:1 variational:3 recently:3 common:1 functional:1 multinomial:1 volume:1 approximates:1 gibbs:4 similarly:1 pq:1 chapelle:3 impressive:1 similarity:1 posterior:6 irrelevant:2 disappointing:1 ubingen:1 inequality:1 qci:1 yi:1 minimum:1 determine:1 aggregated:6 maximize:1 ii:1 afterwards:1 harchaoui:3 technical:3 faster:2 long:1 qi:3 prediction:2 variant:1 regression:10 expectation:3 iteration:3 kernel:7 sometimes:2 invert:1 achieved:4 whereas:5 respondent:3 sch:1 w2:1 rest:2 subject:2 spemannstr:1 legend:1 seem:1 jordan:1 practitioner:1 call:3 wn:1 hb:35 xj:1 gave:1 idea:1 qj:1 utility:12 x1k:6 passing:1 remark:1 useful:1 maybe:1 svms:2 estimated:7 per:2 discrete:1 hyperparameter:1 putting:1 drawn:2 neither:1 immersed:1 sum:1 reasonable:1 guyon:1 decision:1 qui:1 bit:1 bound:1 constraint:3 flat:1 software:6 bibliography:1 conjointly:1 speed:2 span:1 alternate:1 across:1 em:2 wi:20 metropolis:1 explained:1 taken:1 equation:2 turn:2 slack:1 know:1 ordinal:5 available:3 rewritten:1 unreasonable:1 hierarchical:3 include:1 completed:1 marginalized:1 newton:1 unbalance:1 exploit:1 ghahramani:1 build:1 especially:1 micchelli:1 objective:3 question:16 strategy:1 primary:1 traditional:2 diagonal:3 obermayer:1 cw:6 distance:1 simulated:2 barber:1 tuebingen:1 reason:2 consumer:24 assuming:1 length:1 modeled:2 index:1 ratio:3 minimizing:1 providing:2 balance:1 setup:1 trace:2 design:12 implementation:9 anal:1 perform:1 datasets:6 behave:1 heterogeneity:5 incorporated:2 het:2 xpk:4 jebara:1 community:3 rating:6 introduced:1 pair:1 cast:1 extensive:2 optimized:1 boser:1 trans:1 address:2 suggested:1 usually:4 below:2 pattern:1 wi2:2 built:1 max:2 reliable:3 suitable:1 natural:1 rely:1 advanced:1 representing:2 scheme:2 improve:1 hb2:1 tresp:1 text:1 prior:4 l2:3 sawtooth:4 reordering:1 loss:1 fully:2 lecture:1 filtering:2 conjoint:26 xp:1 consistent:1 principle:1 systematically:1 heavy:1 elsewhere:1 surprisingly:1 rasmussen:1 bias:1 allow:1 institute:2 taking:1 boundary:2 evaluating:1 world:2 valid:1 author:1 adaptive:7 far:2 approximate:2 preferred:2 dealing:1 ml:1 global:1 active:6 investigating:1 assumed:1 iterative:1 demanded:1 table:9 additionally:1 promising:1 robust:1 kindly:1 did:2 main:2 rh:3 whole:2 noise:2 hyperparameters:4 profile:4 motivation:1 x1:1 benchmarking:1 simester:1 elaborate:1 cbc:1 tong:1 sub:1 position:1 xl:1 comput:1 crude:1 answering:1 down:4 minute:1 annu:1 svm:4 evidence:1 burden:1 consist:1 workshop:1 vapnik:2 importance:1 magnitude:4 margin:9 simply:1 likely:2 expressed:1 schwaighofer:1 springer:1 refused:1 determines:1 goal:5 identity:1 change:1 specifically:1 hb3:1 hyperplane:1 called:6 total:1 invariance:1 experimental:2 la:1 indicating:1 select:3 college:1 support:4 people:1 kernelization:1 tested:3 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.